Statistical techniques for comparing measurers and methods of measurement: A critical review

被引:360
作者
Ludbrook, J [1 ]
机构
[1] Univ Melbourne, Parkville, Vic 3052, Australia
关键词
categorical variables; continuous variables; correlation; fixed bias; kappa statistic; least products regression analysis; limits of agreement; log-linear modelling; McNemar test; method of differences; proportional bias;
D O I
10.1046/j.1440-1681.2002.03686.x
中图分类号
R9 [药学];
学科分类号
1007 ;
摘要
1. Clinical and experimental pharmacologists and physiologists often wish to compare two methods of measurement, or two measurers. 2. Biostatisticians insist that what should be sought is not agreement between methods or measurers, but disagreement or bias. 3. If measurements have been made on a continuous scale, the main choice is between the Altman-Bland method of differences and least products regression analysis. It is argued that although the former is relatively simple to execute, it does not distinguish adequately between fixed and proportional bias. Least products regression analysis, although more difficult to execute, does achieve this goal. There is almost universal agreement among biostatisticians that the Pearson product-moment correlation coefficient (r ) is valueless as a test for bias. 4. If measurements have been made on a categorical scale, unordered or ordered, the most popular method of analysis is to use the kappa statistic. If the categories are unordered, the unweighted kappa statistic (K) is appropriate. If the categories are ordered, as they are in most rating scales in clinical, psychological and epidemiological research, the weighted kappa statistic (K-w ) is preferable. But K-w corresponds to the intraclass correlation coefficient, which, like r for continuous variables, is incapable of detecting bias. Simple techniques for detecting bias in the case of ordered categorical variables are described and commended to investigators.
引用
收藏
页码:527 / 536
页数:10
相关论文
共 21 条
[1]   MEASUREMENT IN MEDICINE - THE ANALYSIS OF METHOD COMPARISON STUDIES [J].
ALTMAN, DG ;
BLAND, JM .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES D-THE STATISTICIAN, 1983, 32 (03) :307-317
[2]   A NOTE ON THE USE OF THE INTRACLASS CORRELATION-COEFFICIENT IN THE EVALUATION OF AGREEMENT BETWEEN 2 METHODS OF MEASUREMENT [J].
BLAND, JM ;
ALTMAN, DG .
COMPUTERS IN BIOLOGY AND MEDICINE, 1990, 20 (05) :337-340
[3]   Measuring agreement in method comparison studies [J].
Bland, JM ;
Altman, DG .
STATISTICAL METHODS IN MEDICAL RESEARCH, 1999, 8 (02) :135-160
[4]   MEASUREMENTS OF ACUTE CEREBRAL INFARCTION - A CLINICAL EXAMINATION SCALE [J].
BROTT, T ;
ADAMS, HP ;
OLINGER, CP ;
MARLER, JR ;
BARSAN, WG ;
BILLER, J ;
SPILKER, J ;
HOLLERAN, R ;
EBERLE, R ;
HERTZBERG, V ;
RORICK, M ;
MOOMAW, CJ ;
WALKER, M .
STROKE, 1989, 20 (07) :864-870
[5]   BIAS, PREVALENCE AND KAPPA [J].
BYRT, T ;
BISHOP, J ;
CARLIN, JB .
JOURNAL OF CLINICAL EPIDEMIOLOGY, 1993, 46 (05) :423-429
[7]   A COEFFICIENT OF AGREEMENT FOR NOMINAL SCALES [J].
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1960, 20 (01) :37-46
[8]   EQUIVALENCE OF WEIGHTED KAPPA AND INTRACLASS CORRELATION COEFFICIENT AS MEASURES OF RELIABILITY [J].
FLEISS, JL ;
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1973, 33 (03) :613-619
[9]   LARGE SAMPLE STANDARD ERRORS OF KAPPA AND WEIGHTED KAPPA [J].
FLEISS, JL ;
COHEN, J ;
EVERITT, BS .
PSYCHOLOGICAL BULLETIN, 1969, 72 (05) :323-&
[10]   CONFIDENCE-INTERVALS RATHER THAN P-VALUES - ESTIMATION RATHER THAN HYPOTHESIS-TESTING [J].
GARDNER, MJ ;
ALTMAN, DG .
BMJ-BRITISH MEDICAL JOURNAL, 1986, 292 (6522) :746-750