A Systematic Review of the Reliability of Objective Structured Clinical Examination Scores

Document Type

Article

Publication Date

10-2011

Digital Object Identifier (DOI)

https://doi.org/10.1111/j.1365-2923.2011.04075.x

Abstract

Context The objective structured clinical examination (OSCE) is comprised of a series of simulations used to assess the skill of medical practitioners in the diagnosis and treatment of patients. It is often used in high‐stakes examinations and therefore it is important to assess its reliability and validity.

Methods The published literature was searched (PsycINFO, PubMed) for OSCE reliability estimates (coefficient alpha and generalisability coefficients) computed either across stations or across items within stations. Coders independently recorded information about each study. A meta‐analysis of the available literature was computed and sources of systematic variance in estimates were examined.

Results A total of 188 alpha values from 39 studies were coded. The overall (summary) alpha across stations was 0.66 (95% confidence interval [CI] 0.62–0.70); the overall alpha within stations across items was 0.78 (95% CI 0.73–0.82). Better than average reliability was associated with a greater number of stations and a higher number of examiners per station. Interpersonal skills were evaluated less reliably across stations and more reliably within stations compared with clinical skills.

Conclusions Overall scores on the OSCE are often not very reliable. It is more difficult to reliably assess communication skills than clinical skills when considering both as general traits that should apply across situations. It is generally helpful to use two examiners and large numbers of stations, but some OSCEs appear more reliable than others for reasons that are not yet fully understood.

Was this content written or created while at USF?

Yes

Citation / Publisher Attribution

Medical Education, v. 45, issue 12, p. 1181-1189

Share

COinS