TY - GEN
T1 - Human-level Multiple Choice Question Guessing Without Domain Knowledge
T2 - 27th International World Wide Web, WWW 2018
AU - Watson, Patrick
AU - Ma, Teng Fei
AU - Tejwani, Ravi
AU - Chang, Maria
AU - Ahn, Jae Wook
AU - Sundararajan, Sharad
N1 - Publisher Copyright: © 2018 IW3C2 (International World Wide Web Conference Committee), published under Creative Commons CC BY 4.0 License.
PY - 2018/4/23
Y1 - 2018/4/23
N2 - The availability of open educational resources (OER) has enabled educators and researchers to access a variety of learning assessments online. OER communities are particularly useful for gathering multiple choice questions (MCQs), which are easy to grade, but difficult to design well. To account for this, OERs often rely on crowd-sourced data to validate the quality of MCQs. However, because crowds contain many non-experts, and are susceptible to question framing effects, they may produce ratings driven by guessing on the basis of surface-level linguistic features, rather than deep topic knowledge. Consumers of OER multiple choice questions (and authors of original multiple choice questions) would benefit from a tool that automatically provided feedback on assessment quality, and assessed the degree to which OER MCQs are susceptible to framing effects. This paper describes a model that is trained to use domain-naive strategies to guess which multiple choice answer is correct. The extent to which this model can predict the correct answer to an MCQ is an indicator that the MCQ is a poor measure of domain-specific knowledge. We describe an integration of this model with a front-end visualizer and MCQ authoring tool.
AB - The availability of open educational resources (OER) has enabled educators and researchers to access a variety of learning assessments online. OER communities are particularly useful for gathering multiple choice questions (MCQs), which are easy to grade, but difficult to design well. To account for this, OERs often rely on crowd-sourced data to validate the quality of MCQs. However, because crowds contain many non-experts, and are susceptible to question framing effects, they may produce ratings driven by guessing on the basis of surface-level linguistic features, rather than deep topic knowledge. Consumers of OER multiple choice questions (and authors of original multiple choice questions) would benefit from a tool that automatically provided feedback on assessment quality, and assessed the degree to which OER MCQs are susceptible to framing effects. This paper describes a model that is trained to use domain-naive strategies to guess which multiple choice answer is correct. The extent to which this model can predict the correct answer to an MCQ is an indicator that the MCQ is a poor measure of domain-specific knowledge. We describe an integration of this model with a front-end visualizer and MCQ authoring tool.
KW - blind guessing
KW - deep learning
KW - mcqs
KW - oer
UR - https://www.scopus.com/pages/publications/85077979800
U2 - 10.1145/3184558.3186340
DO - 10.1145/3184558.3186340
M3 - Conference contribution
T3 - The Web Conference 2018 - Companion of the World Wide Web Conference, WWW 2018
SP - 299
EP - 303
BT - The Web Conference 2018 - Companion of the World Wide Web Conference, WWW 2018
PB - Association for Computing Machinery, Inc
Y2 - 23 April 2018 through 27 April 2018
ER -