Item response theory facilitated cocalibrating cognitive tests and reduced bias in estimated rates of decline

被引:85
作者
Crane, Paul K. [1 ]
Narasimhalu, Kaavya [1 ]
Gibbons, Laura E. [1 ]
Mungas, Dan A. [2 ]
Haneuse, Sebastien
Larson, Eric B.
Kuller, Lewis [3 ]
Hall, Kathleen [4 ]
van Belle, Gerald [5 ,6 ]
机构
[1] Univ Washington, Dept Med, Seattle, WA 98195 USA
[2] Univ Calif Davis, Dept Neurol, Davis, CA 95616 USA
[3] Univ Pittsburgh, Dept Epidemiol, Pittsburgh, PA 15261 USA
[4] Indiana Univ Purdue Univ, Dept Psychiat, Indianapolis, IN 46202 USA
[5] Univ Washington, Dept Biostat, Seattle, WA 98195 USA
[6] Univ Washington, Dept Environm & Occupat Hlth Sci, Seattle, WA 98195 USA
基金
美国国家卫生研究院;
关键词
cognition; cocalibration; item response theory; psychometrics; longitudinal data analysis; simulation;
D O I
10.1016/j.jclinepi.2007.11.011
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Objective: To cocalibrate the Mini-Mental State Examination, the Modified Mini-Mental State, the Cognitive Abilities Screening Instrument, and the Community Screening Instrument for Dementia using item response theory (IRT) to compare screening cut points used to identify cases of dementia from different studies, to compare measurement properties of the tests, and to explore the implications of these measurement proper-ties on longitudinal studies of cognitive functioning over time. Study Design and Setting: We used cross-sectional data from three large (n > 1000) community-based studies of cognitive functioning in the elderly. We used IRT to cocalibrate the scales and performed simulations of longitudinal studies. Results: Screening cut points varied quite widely across studies. The four tests have curvilinear scaling and varied levels of measurement precision, with more measurement error at higher levels of cognitive functioning. In longitudinal simulations, IRT scores always performed better than standard scoring, whereas a strategy to account for varying measurement precision had mixed results. Conclusion: Cocalibration allows direct comparison of cognitive functioning in studies using any of these four tests. Standard scoring appears to be a poor choice for analysis of longitudinal cognitive testing data. More research is needed into the implications of varying levels of measurement precision. (C) 2008 Elsevier Inc. All rights reserved.
引用
收藏
页码:1018 / 1027
页数:10
相关论文
共 46 条
  • [1] [Anonymous], 2003, Stata Statistical Software
  • [2] [Anonymous], NEW DEV CATEGORICAL
  • [3] Baker F. B., 2004, ITEM RESPONSE THEORY, DOI DOI 10.1201/9781482276725
  • [4] Camilli G., 1994, METHODS IDENTIFYING
  • [5] Differential item functioning analysis with ordinal logistic regression techniques - DIFdetect and difwithpar
    Crane, Paul K.
    Gibbons, Laura E.
    Jolley, Lance
    van Belle, Gerald
    [J]. MEDICAL CARE, 2006, 44 (11) : S115 - S123
  • [6] Differential item functioning related to education and age in the Italian version of the Mini-mental State Examination
    Crane, Paul K.
    Gibbons, Laura E.
    Jolley, Lance
    van Belle, Gerald
    Selleri, Ricardo
    Dalmonte, Edoardo
    De Ronchi, Diana
    [J]. INTERNATIONAL PSYCHOGERIATRICS, 2006, 18 (03) : 505 - 515
  • [7] A 37-item shoulder functional status item pool had negligible differential item functioning
    Crane, PK
    Hart, DL
    Gibbons, LE
    Cook, KF
    [J]. JOURNAL OF CLINICAL EPIDEMIOLOGY, 2006, 59 (05) : 478 - 484
  • [8] Test bias in a cognitive test: differential item functioning in the CASI
    Crane, PK
    van Belle, G
    Larson, EB
    [J]. STATISTICS IN MEDICINE, 2004, 23 (02) : 241 - 256
  • [9] Embretson S. E., 2000, ITEM RESPONSE THEORY, Vxi, P371
  • [10] Incidence and prevalence of dementia in the cardiovascular health study
    Fitzpatrick, AL
    Kuller, LH
    Ives, DG
    Lopez, OL
    Jagust, W
    Breitner, JCS
    Jones, B
    Lyketsos, C
    Dulberg, C
    [J]. JOURNAL OF THE AMERICAN GERIATRICS SOCIETY, 2004, 52 (02) : 195 - 204