Assessing observer agreement when describing and classifying functioning with the international

被引:39
作者
Grill, Eva
Mansmann, Ulrich
Cieza, Alarcos
Stucki, Gerold
机构
[1] Univ Munich, Dept Phys Med & Rehabil, D-81377 Munich, Germany
[2] Univ Munich, Dept Med Informat Iometry & Epidemiol, D-81377 Munich, Germany
[3] Univ Munich, Inst Hlth & Rehabil, WHO FIC Collaborating Ctr DIMDI, ICF Res Branch, D-81377 Munich, Germany
关键词
reproducibility of results; rehabilitation; rater agreement; ICF; log linear models;
D O I
10.2340/16501977-0016
中图分类号
R49 [康复医学];
学科分类号
100215 ;
摘要
Objective: The International Classification of Functioning, Disability and Health (ICF) is used increasingly to describe and classify functioning in medicine without being a psychometrically sound measure. All categories of the ICF are quantified using the same generic 0-4 scale. The objective of this study was to assess observer agreement when describing and classifying functioning with the ICF. Design: A second-level category of the ICF, d430 lifting and carrying objects, was used as an example. To the qualifiers of this category, clinically meaningful definitions were assigned. Data were collected in a cross-sectional survey with repeated measurement. We report raw, specific and chance-corrected measures or agreement, a graphical method and the results of log-linear models for ordinal agreement. Subjects/patients: A convenience sample of patients requiring physical therapy in an acute hospital. Results: Twenty-five patients were assessed twice by 2 observers. Raw agreement was 0.52. Kappa was 0.36, indicating fair agreement. Different hierarchical log-linear models indicated that the strength of agreement was not homogeneous over all categories. Conclusion: Observer agreement has to be evaluated when describing and classifying functioning using the ICF Qualifiers' scale. When assessing inter-observer reliability, the first step is to calculate a summary statistic. Modelling agreement yields valuable insight into the structure of a contingency table, which can lead to further improvement of the scale.
引用
收藏
页码:71 / 76
页数:6
相关论文
共 36 条
[1]   A MODEL FOR AGREEMENT BETWEEN RATINGS ON AN ORDINAL SCALE [J].
AGRESTI, A .
BIOMETRICS, 1988, 44 (02) :539-548
[2]  
[Anonymous], ICF CHECKL VERS 2 1A
[3]  
Bangdiwala S., 1987, P SAS USERS GROUP IN, V12, P1083
[4]   STATISTICAL METHODS FOR ASSESSING AGREEMENT BETWEEN TWO METHODS OF CLINICAL MEASUREMENT [J].
BLAND, JM ;
ALTMAN, DG .
LANCET, 1986, 1 (8476) :307-310
[5]  
Cicchetti D. V., 1971, American Journal of EEG Technology, V11, P101
[6]   Development of ICF core sets for patients with chronic conditions [J].
Cieza, A ;
Ewert, T ;
Üstün, TB ;
Chatterji, S ;
Kostanjsek, N ;
Stucki, G .
JOURNAL OF REHABILITATION MEDICINE, 2004, 36 :9-11
[7]   A COEFFICIENT OF AGREEMENT FOR NOMINAL SCALES [J].
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1960, 20 (01) :37-46
[8]   HIGH AGREEMENT BUT LOW KAPPA .1. THE PROBLEMS OF 2 PARADOXES [J].
FEINSTEIN, AR ;
CICCHETTI, DV .
JOURNAL OF CLINICAL EPIDEMIOLOGY, 1990, 43 (06) :543-549
[9]   EQUIVALENCE OF WEIGHTED KAPPA AND INTRACLASS CORRELATION COEFFICIENT AS MEASURES OF RELIABILITY [J].
FLEISS, JL ;
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1973, 33 (03) :613-619
[10]  
FRIENDLY M, 2000, VISUALISING CATEGORI