Estimating the minimum number of judges required for test-centred standard setting on written assessments. do discussion and iteration have an influence?

Adv Health Sci Educ Theory Pract. 2008 Mar;13(1):11-24. doi: 10.1007/s10459-006-9027-1. Epub 2006 Sep 7.

Abstract

Absolute standard setting procedures are recommended for assessment in medical education. Absolute, test-centred standard setting procedures were introduced for written assessments in the Liverpool MBChB in 2001. The modified Angoff and Ebel methods have been used for short answer question-based and extended matching question-based papers, respectively. Data collected has been analysed to investigate whether reliable standards can be achieved for small-scale, medical school-based assessments, to establish the minimum number of judges required and the effect of a discussion phase on reliability. The root mean squared error (RMSE) has been used as a measure of reliability and used to compute 95% confidence intervals for comparison to the examination statistics. The RMSE has been used to calculate the minimum number of judges required to obtain a predetermined minimum level of reliability, and the effect of the number of judges and number of items have been examined. Values of the RMSE obtained vary from 0.9 to 2.2%. Using average variances across each paper type, the minimum number of judges to obtain a RMSE of less than 2% is 10 or more judges before discussion or 6 or more judges after discussion. The results indicate that including a discussion phase improves the reliability and reduces the minimum number of judges required. Decision studies indicate that increasing the number of questions included in the assessments would not significantly improve the reliability of the standard setting.

MeSH terms

  • Curriculum / standards
  • Education, Medical / standards*
  • Educational Measurement / methods
  • Educational Measurement / standards*
  • England
  • Humans
  • Reproducibility of Results