TY - GEN
T1 - Marking
T2 - 25th International Conference on Artificial Intelligence in Education, AIED 2024
AU - Sonkar, Shashank
AU - Liu, Naiming
AU - Mallick, Debshila B.
AU - Baraniuk, Richard G.
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
PY - 2024
Y1 - 2024
N2 - In this paper, we introduce “Marking”, a novel grading task that enhances automated grading systems by performing an in-depth analysis of student responses and providing students with visual highlights. Unlike traditional systems that provide binary scores, “marking” identifies and categorizes segments of the student response as correct, incorrect, or irrelevant and detects omissions from gold answers. We introduce a new dataset meticulously curated by Subject Matter Experts specifically for this task. We frame “Marking” as an extension of the Natural Language Inference (NLI) task, which is extensively explored in the field of Natural Language Processing. The gold answer and the student response play the roles of premise and hypothesis in NLI, respectively. We subsequently train language models to identify entailment, contradiction, and neutrality from student response, akin to NLI, and with the added dimension of identifying omissions from gold answers. Our experimental setup involves the use of transformer models, specifically BERT and RoBERTa, and an intelligent training step using the e-SNLI dataset. We present extensive baseline results highlighting the complexity of the “Marking” task, which sets a clear trajectory for the upcoming study. Our work not only opens up new avenues for research in AI-powered educational assessment tools, but also provides a valuable benchmark for the AI in education community to engage with and improve upon in the future. The code and dataset can be found at https://github.com/luffycodes/marking.
AB - In this paper, we introduce “Marking”, a novel grading task that enhances automated grading systems by performing an in-depth analysis of student responses and providing students with visual highlights. Unlike traditional systems that provide binary scores, “marking” identifies and categorizes segments of the student response as correct, incorrect, or irrelevant and detects omissions from gold answers. We introduce a new dataset meticulously curated by Subject Matter Experts specifically for this task. We frame “Marking” as an extension of the Natural Language Inference (NLI) task, which is extensively explored in the field of Natural Language Processing. The gold answer and the student response play the roles of premise and hypothesis in NLI, respectively. We subsequently train language models to identify entailment, contradiction, and neutrality from student response, akin to NLI, and with the added dimension of identifying omissions from gold answers. Our experimental setup involves the use of transformer models, specifically BERT and RoBERTa, and an intelligent training step using the e-SNLI dataset. We present extensive baseline results highlighting the complexity of the “Marking” task, which sets a clear trajectory for the upcoming study. Our work not only opens up new avenues for research in AI-powered educational assessment tools, but also provides a valuable benchmark for the AI in education community to engage with and improve upon in the future. The code and dataset can be found at https://github.com/luffycodes/marking.
KW - Automated Grading
KW - Feedback
KW - Highlighting
KW - Natural Language Inference
UR - http://www.scopus.com/inward/record.url?scp=85200210308&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85200210308&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-64302-6_22
DO - 10.1007/978-3-031-64302-6_22
M3 - Conference contribution
AN - SCOPUS:85200210308
SN - 9783031643019
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 309
EP - 323
BT - Artificial Intelligence in Education - 25th International Conference, AIED 2024, Proceedings
A2 - Olney, Andrew M.
A2 - Chounta, Irene-Angelica
A2 - Liu, Zitao
A2 - Santos, Olga C.
A2 - Bittencourt, Ig Ibert
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 8 July 2024 through 12 July 2024
ER -