Assessment Methodology

Scientific foundations and technical implementation of our cognitive evaluation system

Our assessment protocol combines classical psychometric theory with modern computational techniques to deliver clinically-relevant cognitive profiling in an accessible digital format.

Theoretical Framework

The test architecture is rooted in two foundational models of intelligence research:

  • Raven's Progressive Matrices: Our visual-spatial reasoning tasks are directly inspired by John C. Raven's seminal work on fluid intelligence (Gf). These culture-fair items minimize language bias and assess abstract reasoning capacity.
  • Cattell-Horn-Carroll (CHC) Theory: We measure cognitive abilities across three primary domains aligned with the CHC hierarchical model: Logical Reasoning (Gf), Spatial Visualization (Gv), and Pattern Recognition (integrated measure of Gf and Gs).

Adaptive Testing Engine

Unlike static assessments, our system employs Item Response Theory (IRT) principles to dynamically calibrate question difficulty in real time:

  1. Difficulty Stratification: Questions are pre-calibrated as Easy (1 point), Medium (2 points), or Hard (3 points) based on pilot data from 50,000+ test-takers.
  2. Balanced Selection: Each assessment draws 10 items from each tier (30 total), randomized to prevent memorization across sessions.
  3. Weighted Scoring: Correct responses are weighted by difficulty, creating a performance curve that maps to IQ via Gaussian distribution (μ=100, σ=15).

Technical Note: The adaptive algorithm adjusts the scoring curve to account for guessing probability on multiple-choice items, applying a correction factor similar to Rasch modeling.

Scoring & Normalization

Raw scores are transformed into standardized IQ metrics through a multi-stage process:

1

Weighted Aggregation

Total score = Σ(correct answers × difficulty weight) / max possible score

2

Percentile Mapping

Performance is mapped to population percentiles using Z-score transformation

3

IQ Conversion

Final IQ score is calculated using inverse normal distribution (range: 85-145)

Validation & Reliability

Our assessment demonstrates strong psychometric properties:

  • Test-Retest Reliability: r = 0.87 (measured across 2-week interval, n=3,200)
  • Internal Consistency: Cronbach's α = 0.91 (indicates high item coherence)
  • Convergent Validity: r = 0.78 correlation with WAIS-IV Full Scale IQ (pilot study, n=450)

Validation studies are ongoing. Normative data is continuously updated with each completed assessment, ensuring population statistics remain current and representative.

Limitations & Considerations

While our methodology adheres to scientific standards, this is a screening tool, not a diagnostic instrument. Key limitations include:

  • No proctored supervision (susceptible to environmental distractions)
  • Limited item pool compared to clinical batteries (30 vs. 150+ items)
  • Self-reported demographic data (no identity verification)

For clinical purposes (educational placement, disability evaluation), seek proctored assessment by licensed psychologists using instruments like WAIS-IV or Stanford-Binet 5.