Standard setting in an objective structured clinical examination: use of global ratings of borderline performance to determine the passing score

被引:64
作者
Wilkinson, T
Newble, DI
Frampton, CM
机构
[1] Univ Otago, Christchurch Sch Med, Christchurch, New Zealand
[2] Univ Sheffield, Dept Med Educ, Sheffield, S Yorkshire, England
关键词
clinical competence; education; medical; undergraduate; methods; educational measurement; New Zealand; reproducibility of results;
D O I
10.1046/j.1365-2923.2001.01041.x
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Background Objective structured clinical examination (OSCE) standard-setting procedures are not well developed and are often time-consuming and complex. We report an evaluation of a simple 'contrasting groups' method, applied to an OSCE conducted simultaneously in three separate schools. Subjects Medical students undertaking an end-of-fifth year multidisciplinary OSCE. Methods Using structured marking sheets, pairs of examiners independently scored student performance at each OSCE station. Examiners also provided a global rating of overall performance. The actual scores of any borderline candidates at each station were averaged to provide a passing score for each station. The passing scores for all stations were combined to become the passing score for the whole exam. Validity was determined by making comparisons with performance on other fifth-year assessments. Reliability measures comprised interschool agreement, interexaminer agreement and interstation variability. Results The approach was simple and had face validity. There was a stronger association between the performance of borderline candidates on the OSCE and their in-course assessments than with their performance on the written exam, giving a weak measure of construct validity in the absence of a better 'gold standard'. There was good agreement between examiners in identifying borderline candidates. There were significant differences between schools in the borderline score for some stations, which disappeared when more than three stations were aggregated. Conclusion This practical method provided a valid and reliable competence-based pass mark. Combining marks from all stations before determining the pass mark was more reliable than making decisions based on individual stations.
引用
收藏
页码:1043 / 1049
页数:7
相关论文
共 6 条
[1]   AMEE Guide No. 18: Standard setting in student assessment [J].
Ben-David, MF .
MEDICAL TEACHER, 2000, 22 (02) :120-130
[2]   The passing score in the objective structured clinical examination [J].
Morrison, H ;
McNally, H ;
Wylie, C ;
McFaul, P ;
Thompson, W .
MEDICAL EDUCATION, 1996, 30 (05) :345-348
[3]  
Newble D., 1994, Teach. Learn. Med. Int. J, V6, P213, DOI [10.1080/10401339409539680, DOI 10.1080/10401339409539680]
[4]   Standards and reliability in evaluation: When rules of thumb don't apply [J].
Norcini, JJ .
ACADEMIC MEDICINE, 1999, 74 (10) :1088-1090
[5]  
ROTHMAN AI, 1997, ADV HEALTH SCI EDUC, V1, P215
[6]   Development of a three-centre simultaneous objective structured clinical examination [J].
Wilkinson, TJ ;
Newble, DI ;
Wilson, PD ;
Carter, JM ;
Helms, RM .
MEDICAL EDUCATION, 2000, 34 (10) :798-807