THE INTERRATER RELIABILITY AND INTERNAL CONSISTENCY OF A CLINICAL-EVALUATION EXERCISE

被引:60
作者
KROBOTH, FJ
HANUSA, BH
PARKER, S
COULEHAN, JL
KAPOOR, WN
BROWN, FH
KARPF, M
LEVEY, GS
机构
[1] University of Pittsburgh, Internal Medicine, Pittsburgh, 15261, PA, 190 Lothrop Street
关键词
CLINICAL EVALUATION EXERCISE; INTERRATER RELIABILITY; EDUCATION; PERFORMANCE ASSESSMENT;
D O I
10.1007/BF02598008
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Objective: To assess the internal consistency and inter-rater reliability of a clinical evaluation exercise (CEX) format that was designed to be easily utilized, but sufficiently detailed, to achieve uniform recording of the observed examination. Design: A comparison of 128 CEXs conducted for 32 internal medicine interns by full-time faculty. This paper reports alpha coefficients as measures of internal consistency and several measures of inter-rater reliability. Setting: A university internal medicine program. Observations were conducted at the end of the internship year. Participants: Participants were 32 interns and observers were 12 full-time faculty in the department of medicine. The entire intern group was chosen in order to optimize the spectrum of abilities represented. Patients used for the study were recruited by the chief resident from the inpatient medical service based on their ability and willingness to participate. Intervention: Each intern was observed twice and there were two examiners during each CEX. The examiners were given a standardized preparation and used a format developed over five years of previous pilot studies. Measurements and main results: The format appeared to have excellent internal consistency; alpha coefficients ranged from 0.79 to 0.99. However, multiple methods of determining inter-rater reliability yielded similar results; intraclass correlations ranged from 0.23 to 0.50 and generalizability coefficients from a low of 0.00 for the overall rating of the CEX to a high of 0.61 for the physical examination section. Transforming scores to eliminate rater effects and dichotomizing results into pass-fail did not appear to enhance the reliability results. Conclusions: Although the CEX is a valuable didactic tool, its psychometric properties preclude reliable assessment of clinical skills as a one-time observation.
引用
收藏
页码:174 / 179
页数:6
相关论文
共 27 条
[1]  
Blank L.L., Grosso L.J., Benson J.A., A survey of clinical skills evaluation practices in internal medicine residency programs, J Med Educ, 59, pp. 401-6, (1984)
[2]  
Petersdorf R.G., Beck J.C., The new procedure for evaluating the clinical competence of candidates to be certified by the American Board of Internal Medicine, Ann Intern Med, 76, pp. 491-6, (1972)
[3]  
Woolliscroft J.O., Stross J.K., Silva J., Clinical competence certification: a critical appraisal, J Med Educ, 59, pp. 799-805, (1984)
[4]  
Herbers J.E., Noel G.L., Cooper G.S., Harvey J., Pangaro L.N., Weaver M.J., How accurate are faculty evaluations of clinical competence?, J Gen Intern Med, 4, pp. 202-8, (1989)
[5]  
Kroboth F.J., Kapoor W., Brown F.H., Karpf M., Levey G.S., A comparative trial of the clinical evaluation exercise, Arch Intern Med, 145, pp. 1121-3, (1985)
[6]  
Lipkin M., The medical interview and related skills, Office practice of medicine, pp. 1287-306, (1987)
[7]  
Brennan R.L., Kane M.T., Generalizability theory: a review, New directions for testing and measurement (no. 4): methodological developments, pp. 33-51, (1979)
[8]  
Shrout P.E., Fleiss J.L., Intraclass correlations in assessing rater reliability, Psychol Bull, 86, pp. 420-8, (1979)
[9]  
Cohen J., Weighted kappa: nominal scale agreement with provisions for scaled disagreement or partial credit, Psychol Bull, 70, pp. 213-20, (1968)
[10]  
Cohen J., A co-efficient of agreement for nominal scales, Educational and Psychological Measurement, 20, pp. 37-46, (1960)