eCite Digital Repository

Inter-rater reliability: comparison of checklist and global scoring for OSCEs


Malau-Aduli, BS and Mulcahy, S and Warnecke, E and Otahal, P and Teague, PA and Turner, R and Van der Vleuten, C, Inter-rater reliability: comparison of checklist and global scoring for OSCEs, Creative Education, 3, (6A) pp. 937-942. ISSN 2151-4755 (2012) [Refereed Article]

Restricted - Request a copy

Copyright Statement

Copyright 2012 SciRes

Official URL:

DOI: doi:10.4236/ce.2012.326142


Objective Structured Clinical Examinations (OSCEs) have been used globally in evaluating clinical com- petence in the education of health professionals. Despite the objective intent of OSCEs, scoring methods used by examiners have been a potential source of measurement error affecting the precision with which test scores are determined. In this study, we investigated the differences in the inter-rater reliabilities of objective checklist and subjective global rating scores of examiners (who were exposed to an online training program to standardise scoring techniques) across two medical schools. Examiners’ perceptions of the e-scoring program were also investigated. Two Australian universities shared three OSCE stations in their end-of-year undergraduate medical OSCEs. The scenarios were video-taped and used for on-line examiner training prior to actual exams. Examiner ratings of performance at both sites were analysed using generalisability theory. A single facet, all random persons by raters design [PxR] was used to measure inter-rater reliability for each station, separate for checklist scores and global ratings. The resulting variance components were pooled across stations and examination sites. Decision studies were used to meas- ure reliability estimates. There was no significant mean score difference between examination sites. Variation in examinee ability accounted for 68.3% of the total variance in checklist scores and 90.2% in global ratings. Rater contribution was 1.4% & 0% of the total variance in checklist score and global rating respectively, reflecting high inter-rater reliability of the scores provided by co-examiners across the two schools. Score variance due to interaction and residual error was larger for checklist scores (30.3% vs 9.7%) than for global ratings. Reproducibility coefficients for global ratings were higher than for checklist scores. Survey results showed that the e-scoring package facilitated consensus on scoring techniques. This approach to examiner training also allowed examiners to calibrate the OSCEs in their own time. This study revealed that inter-rater reliability was higher for global ratings than for checklist scores, thus providing further evidence for the reliability of subjective examiner ratings.

Item Details

Item Type:Refereed Article
Research Division:Education
Research Group:Specialist studies in education
Research Field:Education assessment and evaluation
Objective Division:Education and Training
Objective Group:Teaching and curriculum
Objective Field:Assessment, development and evaluation of curriculum
UTAS Author:Malau-Aduli, BS (Dr Bunmi Malau-Aduli)
UTAS Author:Mulcahy, S (Ms Sue Mulcahy)
UTAS Author:Warnecke, E (Dr Emma Warnecke)
UTAS Author:Otahal, P (Mr Petr Otahal)
UTAS Author:Turner, R (Professor Richard Turner)
ID Code:80889
Year Published:2012
Deposited By:Medicine
Deposited On:2012-11-15
Last Modified:2014-09-23
Downloads:2 View Download Statistics

Repository Staff Only: item control page