The reliability of workplace-based assessment in postgraduate medical education and training: a national evaluation in general practice in the United Kingdom

Adv Health Sci Educ Theory Pract. 2009 May;14(2):219-32. doi: 10.1007/s10459-008-9104-8. Epub 2008 Feb 28.

Abstract

To investigate the reliability and feasibility of six potential workplace-based assessment methods in general practice training: criterion audit, multi-source feedback from clinical and non-clinical colleagues, patient feedback (the CARE Measure), referral letters, significant event analysis, and video analysis of consultations. Performance of GP registrars (trainees) was evaluated with each tool to assess the reliabilities of the tools and feasibility, given raters and number of assessments needed. Participant experience of process determined by questionnaire. 171 GP registrars and their trainers, drawn from nine deaneries (representing all four countries in the UK), participated. The ability of each tool to differentiate between doctors (reliability) was assessed using generalisability theory. Decision studies were then conducted to determine the number of observations required to achieve an acceptably high reliability for "high-stakes assessment" using each instrument. Finally, descriptive statistics were used to summarise participants' ratings of their experience using these tools. Multi-source feedback from colleagues and patient feedback on consultations emerged as the two methods most likely to offer a reliable and feasible opinion of workplace performance. Reliability co-efficients of 0.8 were attainable with 41 CARE Measure patient questionnaires and six clinical and/or five non-clinical colleagues per doctor when assessed on two occasions. For the other four methods tested, 10 or more assessors were required per doctor in order to achieve a reliable assessment, making the feasibility of their use in high-stakes assessment extremely low. Participant feedback did not raise any major concerns regarding the acceptability, feasibility, or educational impact of the tools. The combination of patient and colleague views of doctors' performance, coupled with reliable competence measures, may offer a suitable evidence-base on which to monitor progress and completion of doctors' training in general practice.

Publication types

  • Validation Study

MeSH terms

  • Adult
  • Clinical Competence / standards
  • Clinical Competence / statistics & numerical data*
  • Educational Measurement
  • Educational Status
  • Family Practice / education
  • Family Practice / standards
  • Family Practice / statistics & numerical data*
  • Feasibility Studies
  • Female
  • Humans
  • Internship and Residency / standards
  • Internship and Residency / statistics & numerical data*
  • Male
  • Patient Satisfaction
  • Pilot Projects
  • Reproducibility of Results
  • Surveys and Questionnaires
  • United Kingdom
  • Workplace / standards
  • Workplace / statistics & numerical data*