Interobserver agreement issues in radiology

被引:252
作者
Benchoufi, M. [1 ,2 ]
Matzner-Lober, E. [3 ]
Molinari, N. [4 ]
Jannot, A-S [2 ,5 ,6 ]
Soyer, P. [2 ,7 ]
机构
[1] Natl Inst Agr Res INRA, French Inst Hlth & Med Res Inserm, Ctr Res Epidemiol & Stat CRESS, Paris, France
[2] Univ Paris, F-75006 Paris, France
[3] CREST ENSAE, UMR 9194, F-91120 Palaiseau, France
[4] Univ Montpellier, Dept Stat, CHU Montpellier, CNRS,IMAG, F-34000 Montpellier, France
[5] Sorbonne Univ, Ctr Rech Cordeliers, INSERM, F-75006 Paris, France
[6] Hop Europeen Georges Pompidou, Assistance Publ Hop Paris AP HP, Dept Biostat Med Informat & Publ Hlth, F-75015 Paris, France
[7] Hop Cochin, Assistance Publ Hop Paris AP HP, Dept Radiol, F-75014 Paris, France
关键词
Reproducibility of results; Interobserver agreement; Radiology; Kappa test; Intraclass correlation coefficient; BLAND-ALTMAN; STATISTICAL-METHODS; KAPPA; RELIABILITY;
D O I
10.1016/j.diii.2020.09.001
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
100231 [临床病理学]; 100902 [航空航天医学];
摘要
Agreement between observers (i.e., inter-rater agreement) can be quantified with various criteria but their appropriate selections are critical. When the measure is qualitative (nominal or ordinal), the proportion of agreement or the kappa coefficient should be used to evaluate inter-rater consistency (i.e., inter-rater reliability). The kappa coefficient is more meaningful that the raw percentage of agreement, because the latter does not account for agreements due to chance alone. When the measures are quantitative, the intraclass correlation coefficient (ICC) should be used to assess agreement but this should be done with care because there are different ICCs so that it is important to describe the model and type of ICC being used. The Bland-Altman method can be used to assess consistency and conformity but its use should be restricted to comparison of two raters. (C) 2020 Societe francaise de radiologie. Published by Elsevier Masson SAS. All rights reserved.
引用
收藏
页码:639 / 641
页数:3
相关论文
共 22 条
[1]
Consensus Interpretation in Imaging Research: Is There a Better Way? [J].
Bankier, Alexander A. ;
Levine, Deborah ;
Halpern, Elkan F. ;
Kressel, Herbert Y. .
RADIOLOGY, 2010, 257 (01) :14-17
[2]
STATISTICAL METHODS FOR ASSESSING AGREEMENT BETWEEN TWO METHODS OF CLINICAL MEASUREMENT [J].
BLAND, JM ;
ALTMAN, DG .
LANCET, 1986, 1 (8476) :307-310
[3]
Dynamic contrast-enhanced MR imaging o the rectum: Correlations between single-section and whole-tumor histogram analyses [J].
Choi, M. H. ;
Oh, S. N. ;
Park, G. E. ;
Yeo, D-M ;
Jung, S. E. .
DIAGNOSTIC AND INTERVENTIONAL IMAGING, 2018, 99 (09) :537-545
[4]
Bland-Altman analysis: A paradigm to understand correlation and agreement [J].
Dogan, Nurettin Ozgur .
TURKISH JOURNAL OF EMERGENCY MEDICINE, 2018, 18 (04) :139-141
[5]
HIGH AGREEMENT BUT LOW KAPPA .1. THE PROBLEMS OF 2 PARADOXES [J].
FEINSTEIN, AR ;
CICCHETTI, DV .
JOURNAL OF CLINICAL EPIDEMIOLOGY, 1990, 43 (06) :543-549
[6]
Fleiss JL, PSYCHOL BULL1971, V76, P378
[7]
Statistical methods for assessing agreement for ordinal data [J].
Jakobsson, U ;
Westergren, A .
SCANDINAVIAN JOURNAL OF CARING SCIENCES, 2005, 19 (04) :427-431
[8]
A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research [J].
Koo, Terry K. ;
Li, Mae Y. .
JOURNAL OF CHIROPRACTIC MEDICINE, 2016, 15 (02) :155-163
[9]
MEASUREMENT OF OBSERVER AGREEMENT FOR CATEGORICAL DATA [J].
LANDIS, JR ;
KOCH, GG .
BIOMETRICS, 1977, 33 (01) :159-174
[10]
Intra peritoneal abdominal fat area measured from computed tomography is an independent factor of severe acute pancreatitis [J].
Madico, C. ;
Herpe, G. ;
Vesselle, G. ;
Boucebci, S. ;
Tougeron, D. ;
Sylvain, C. ;
Ingrand, P. ;
Tasu, J-P .
DIAGNOSTIC AND INTERVENTIONAL IMAGING, 2019, 100 (7-8) :421-426