摘要:Production of high-quality multiple-choice questions (MCQs) for both formative and
summative assessments is a time-consuming task requiring great skill, creativity and
insight. The transition to online examinations, with the concomitant exposure of previously tried-and-tested MCQs, exacerbates the challenges of question production and
highlights the need for innovative solutions. Several groups have shown that it is practical to leverage the student cohort to produce a very large number of syllabus-aligned
MCQs for study banks. Although student-generated questions are well suited for formative feedback and practice activities, they are generally not thought to be suitable
for high-stakes assessments. In this study, we aimed to demonstrate that training can
be provided to students in a scalable fashion to generate questions of similar quality to
those produced by experts and that identification of suitable questions can be achieved
with minimal academic review and editing. Second-year biochemistry and molecular
biology students were assigned a series of activities designed to coach them in the art
of writing and critiquing MCQs. This training resulted in the production of over 1000
MCQs that were then gauged for potential by either expert academic judgement or via
a data-driven approach in which the questions were trialled objectively in a low-stakes
test. Questions selected by either method were then deployed in a high-stakes in-semester assessment alongside questions from two academically authored sources: textbook-derived MCQs and past paper questions. A total of 120 MCQs from these four
sources were deployed in assessments attempted by over 600 students. Each question
was subjected to rigorous performance analysis, including the calculation of standard
metrics from classical test theory and more sophisticated item response theory (IRT)
measures. The results showed that MCQs authored by students, and selected at low
cost, performed as well as questions authored by academics, illustrating the potential
of this strategy for the efficient creation of large numbers of high-quality MCQs for
summative assessment.
其他摘要:Production of high-quality multiple-choice questions (MCQs) for both formative and summative assessments is a time-consuming task requiring great skill, creativity and insight. The transition to online examinations, with the concomitant exposure of previously tried-and-tested MCQs, exacerbates the challenges of question production and highlights the need for innovative solutions. Several groups have shown that it is practical to leverage the student cohort to produce a very large number of syllabus-aligned MCQs for study banks. Although student-generated questions are well suited for formative feedback and practice activities, they are generally not thought to be suitable for high-stakes assessments. In this study, we aimed to demonstrate that training can be provided to students in a scalable fashion to generate questions of similar quality to those produced by experts and that identification of suitable questions can be achieved with minimal academic review and editing. Second-year biochemistry and molecular biology students were assigned a series of activities designed to coach them in the art of writing and critiquing MCQs. This training resulted in the production of over 1000 MCQs that were then gauged for potential by either expert academic judgement or via a data-driven approach in which the questions were trialled objectively in a low-stakes test. Questions selected by either method were then deployed in a high-stakes in-semester assessment alongside questions from two academically authored sources: textbook-derived MCQs and past paper questions. A total of 120 MCQs from these four sources were deployed in assessments attempted by over 600 students. Each question was subjected to rigorous performance analysis, including the calculation of standard metrics from classical test theory and more sophisticated item response theory (IRT) measures. The results showed that MCQs authored by students, and selected at low cost, performed as well as questions authored by academics, illustrating the potential of this strategy for the efficient creation of large numbers of high-quality MCQs for summative assessment.