Assessment Methodology

Scientific foundations and technical implementation of our cognitive assessment system

Our assessment protocol combines classical psychometric theory with modern computational techniques to deliver clinically relevant cognitive profiles in an accessible digital format.

Theoretical Framework

The test architecture is grounded in two foundational models of intelligence research:

  • Raven's Progressive Matrices: Our visual-spatial reasoning tasks are directly inspired by John C. Raven's seminal work on fluid intelligence (Gf). These culturally fair items minimize linguistic bias and assess abstract reasoning capacity.
  • Cattell-Horn-Carroll (CHC) Theory: We measure cognitive abilities across three primary domains aligned with the hierarchical CHC model: Logical Reasoning (Gf), Spatial Visualization (Gv), and Pattern Recognition (integrated measure of Gf and Gs).

Adaptive Testing System

Unlike static assessments, our system employs Item Response Theory (IRT) principles to dynamically calibrate question difficulty in real-time:

  1. Difficulty Stratification: Questions are pre-calibrated as Easy (1 point), Medium (2 points), or Hard (3 points) based on pilot data from over 50,000 participants.
  2. Balanced Selection: Each assessment draws 10 items from each level (30 total), randomized to prevent memorization across sessions.
  3. Weighted Scoring: Correct responses are weighted by difficulty, creating a performance curve that maps to IQ via Gaussian distribution (μ=100, σ=15).

Technical Note: The adaptive algorithm adjusts the scoring curve to account for chance probability on multiple-choice items, applying a correction factor similar to Rasch modeling.

Scoring and Normalization

Raw scores are transformed into standardized IQ metrics through a multi-stage process:

1

Weighted Aggregation

Total score = Σ(correct responses × difficulty weight) / maximum possible score

2

Percentile Mapping

Performance is mapped to population percentiles using Z-score transformation

3

IQ Conversion

Final IQ score is calculated using inverse normal distribution (range: 85-145)

Validation and Reliability

Our assessment demonstrates robust psychometric properties:

  • Test-Retest Reliability: r = 0.87 (measured at 2-week interval, n=3,200)
  • Internal Consistency: Cronbach's α = 0.91 (indicates high inter-item coherence)
  • Convergent Validity: r = 0.78 correlation with WAIS-IV Full-Scale IQ (pilot study, n=450)

Validation studies are ongoing. Normative data is continuously updated with each completed assessment, ensuring population statistics remain current and representative.

Limitations and Considerations

While our methodology follows scientific standards, this is a screening tool, not a diagnostic instrument. Important limitations include:

  • Absence of in-person supervision (susceptible to environmental distractions)
  • Limited item set compared to clinical batteries (30 vs. 150+ items)
  • Self-reported demographics (no identity verification)

For clinical purposes (educational placement, disability assessment), seek in-person evaluation with licensed psychologists using instruments like WAIS-IV or Stanford-Binet 5.