Variation between experienced observers in the interpretation of accident and emergency radiographs

被引:119
作者
Robinson, PJA [1 ]
Wilson, D
Coral, A
Murphy, A
Verow, P
机构
[1] St Jamess Univ Hosp, Clin Radiol Dept, Leeds, W Yorkshire, England
[2] St Jamess Univ Hosp, Med Phys Unit, Leeds, W Yorkshire, England
[3] York Dist Hosp, Dept Radiol, York, N Yorkshire, England
关键词
D O I
10.1259/bjr.72.856.10474490
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Skill mix and role extension initiatives have highlighted the difficulty of establishing quality standards for the accuracy of plain film reporting. An acceptable performance might be one which is indistinguishable from that of a group of experienced consultant radiologists. In order to assess the feasibility of setting such a standard, the variation between experienced observers must first be established. This study examines the variation found between three observers with the three major types of plain film examination. 402 plain film examinations (205 skeletal, 100 chest and 97 abdominal) performed on accident and emergency patients were reported retrospectively and independently by three experienced radiologists. The clinical data supplied on the request cards were available to the readers. Each examination was categorized by each reader as being normal, as showing significant abnormality relevant to the current clinical problem, or as showing insignificant or irrelevant abnormality. Concordance between all three readers was found in 51%, 61% and 74% of abdominal, chest and skeletal radiographs, respectively. Weighted kappa values confirmed that the level of agreement between pairs of observers was higher with skeletal radiographs (K-w = 0.76-0.77) than with chest (K-w,= 0.63-0.68), or abdominal (K-w = 0.50-0.78) examinations. However, the frequency of major disagreements (at least one reader reporting "normal" and one reporting "relevant abnormality") was similar for abdominal (11%), chest (12%) and skeletal (10%) radiographs. When the reports were reclassified into only two groups-either significantly abnormal or not-pairs of observers disagreed on 9-10% of skeletal, 11-19% of chest and 8-18% of abdominal cases. The average incidence of errors per observer was estimated to be in the range 3-6%. The magnitude of interobserver variation in plain film reporting is considerable, and must be taken into account when designing assessment techniques and setting quality standards for this activity.
引用
收藏
页码:323 / 330
页数:8
相关论文
共 38 条
[1]  
*AM COLL RAD, 1995, ACR STAND COMM DIAGN
[2]  
AYERS B, 1997, BR J RADIOL S, V70, P36
[3]   MALPRACTICE AND RADIOLOGISTS IN COOK-COUNTY, IL - TRENDS IN 20 YEARS OF LITIGATION [J].
BERLIN, L ;
BERLIN, JW .
AMERICAN JOURNAL OF ROENTGENOLOGY, 1995, 165 (04) :781-788
[4]   CHEST ROENTGENOGRAMS IN DIAGNOSIS OF TRAUMATIC RUPTURE OF THE AORTA - OBSERVER VARIATION IN INTERPRETATION [J].
BURNEY, RE ;
GUNDRY, SR ;
MACKENZIE, JR ;
WHITEHOUSE, WM ;
WU, SC .
CHEST, 1984, 85 (05) :605-609
[5]   OBSERVER VARIATION IN DETECTING THE RADIOLOGIC FEATURES ASSOCIATED WITH BRONCHIOLITIS [J].
COBLENTZ, CL ;
BABCOOK, CJ ;
ALTON, D ;
RILEY, BJ ;
NORMAN, G .
INVESTIGATIVE RADIOLOGY, 1991, 26 (02) :115-118
[6]   DETECTION OF SIGNIFICANT ABNORMALITIES ON LUMBAR SPINE RADIOGRAPHS [J].
DAVIES, AM ;
FOWLER, J ;
TYRRELL, PNM ;
MILLAR, JS ;
LEAHY, JF ;
PATEL, K ;
HILL, JS .
BRITISH JOURNAL OF RADIOLOGY, 1993, 66 (781) :37-43
[7]   The statistical analysis of kappa statistics in multiple samples [J].
Donner, A ;
Klar, N .
JOURNAL OF CLINICAL EPIDEMIOLOGY, 1996, 49 (09) :1053-1058
[8]  
Ekstrom G, 1993, J Forensic Odontostomatol, V11, P45
[9]   EQUIVALENCE OF WEIGHTED KAPPA AND INTRACLASS CORRELATION COEFFICIENT AS MEASURES OF RELIABILITY [J].
FLEISS, JL ;
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1973, 33 (03) :613-619