Am I Wrong, or Is the AutograderWrong? Effects of AI Grading Mistakes on Learning

PROCEEDINGS OF THE 2023 ACM CONFERENCE ON INTERNATIONAL COMPUTING EDUCATION RESEARCH V.1, ICER 2023 V1(2023)

引用 0|浏览4
暂无评分
摘要
Errors in AI grading and feedback often have an intractable set of causes and are, by their nature, difficult to completely avoid. Since inaccurate feedback potentially harms learning, there is a need for designs and workflows that mitigate these harms. To better understand the mechanisms by which erroneous AI feedback impacts students' learning, we conducted surveys and interviews that recorded students' interactions with a short-answer AI autograder for "Explain in Plain English" code reading problems. Using causal modeling, we inferred the learning impacts of wrong answers marked as right (false positives, FPs) and right answers marked as wrong (false negatives, FNs). We further explored explanations for the learning impacts, including errors influencing participants' engagement with feedback and assessments of their answers' correctness, and participants' prior performance in the class. FPs harmed learning in large part due to participants' failures to detect the errors. This was due to participants not paying attention to the feedback after being marked as right, and an apparent bias against admitting one's answer was wrong once marked right. On the other hand, FNs harmed learning only for survey participants, suggesting that interviewees' greater behavioral and cognitive engagement protected them from learning harms. Based on these findings, we propose ways to help learners detect FPs and encourage deeper reflection on FNs to mitigate the learning harms of AI errors.
更多
查看译文
关键词
human-AI interaction,AI error,formative feedback,autograder,computer science education,automated short answer grading,explain in plain English,EiPE,Bayesian modeling
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要