Item response theory facilitated cocalibrating cognitive tests and reduced bias in estimated rates of decline

J Clin Epidemiol. 2008 Oct;61(10):1018-27.e9. doi: 10.1016/j.jclinepi.2007.11.011. Epub 2008 May 5.

Abstract

Objective: To cocalibrate the Mini-Mental State Examination, the Modified Mini-Mental State, the Cognitive Abilities Screening Instrument, and the Community Screening Instrument for Dementia using item response theory (IRT) to compare screening cut points used to identify cases of dementia from different studies, to compare measurement properties of the tests, and to explore the implications of these measurement properties on longitudinal studies of cognitive functioning over time.

Study design and setting: We used cross-sectional data from three large (n>1000) community-based studies of cognitive functioning in the elderly. We used IRT to cocalibrate the scales and performed simulations of longitudinal studies.

Results: Screening cut points varied quite widely across studies. The four tests have curvilinear scaling and varied levels of measurement precision, with more measurement error at higher levels of cognitive functioning. In longitudinal simulations, IRT scores always performed better than standard scoring, whereas a strategy to account for varying measurement precision had mixed results.

Conclusion: Cocalibration allows direct comparison of cognitive functioning in studies using any of these four tests. Standard scoring appears to be a poor choice for analysis of longitudinal cognitive testing data. More research is needed into the implications of varying levels of measurement precision.

Publication types

  • Meta-Analysis
  • Research Support, N.I.H., Extramural
  • Research Support, Non-U.S. Gov't
  • Review

MeSH terms

  • Aged
  • Aged, 80 and over
  • Cognition Disorders / diagnosis*
  • Cognition Disorders / etiology
  • Dementia / diagnosis*
  • Dementia / psychology
  • Disease Progression
  • Epidemiologic Methods
  • Female
  • Humans
  • Male
  • Neuropsychological Tests / standards*
  • Psychometrics

Grants and funding