| dc.contributor.author | Ahmad, Mak | |
| dc.contributor.author | Ravi, Prerna | |
| dc.contributor.author | Karger, David | |
| dc.contributor.author | Facciotti, Marc | |
| dc.date.accessioned | 2025-08-28T20:59:14Z | |
| dc.date.available | 2025-08-28T20:59:14Z | |
| dc.date.issued | 2025-07-17 | |
| dc.identifier.isbn | 979-8-4007-1291-3 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/162577 | |
| dc.description | L@S ’25, Palermo, Italy | en_US |
| dc.description.abstract | Providing personalized, detailed feedback at scale in large undergraduate STEM courses remains a persistent challenge. We present
an empirically evaluated practice exam system that integrates AI
generated feedback with targeted textbook references, deployed in
a large introductory biology course. Our system specifically aims
to encourage metacognitive behavior by asking students to explain
their answers and declare their confidence. It uses OpenAI’s GPT4o to generate personalized feedback based on this information,
while directing them to relevant textbook sections. Through detailed interaction logs from consenting participants across three
midterms (541, 342, and 413 students respectively), totaling 28,313
question-student interactions across 146 learning objectives, along
with 279 post-exam surveys and 23 semi-structured interviews, we
examined the system’s impact on learning outcomes and student
engagement. Analysis showed that across all midterms, the different feedback types showed no statistically significant differences in
performance, though there were some trends suggesting potential
benefits worth further investigation. The system’s most substantial impact emerged through its required confidence ratings and
explanations, which students reported transferring to their actual
exam strategies. Approximately 40% of students engaged with textbook references when prompted by feedback—significantly higher
than traditional reading compliance rates. Survey data revealed
high student satisfaction (M=4.1/5), with 82.1% reporting increased
confidence on midterm topics they had practiced, and 73.4% indicating they could recall and apply specific concepts from practice
sessions. Our findings demonstrate how thoughtfully designed AIenhanced systems can scale formative assessment while promoting
sustainable study practices and self-regulated learning behaviors,
suggesting that embedding structured reflection requirements may
be more impactful than sophisticated feedback mechanisms. | en_US |
| dc.publisher | ACM|Proceedings of the Twelfth ACM Conference on Learning @ Scale | en_US |
| dc.relation.isversionof | https://doi.org/10.1145/3698205.3729542 | en_US |
| dc.rights | Creative Commons Attribution | en_US |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | en_US |
| dc.source | Association for Computing Machinery | en_US |
| dc.title | How Adding Metacognitive Requirements in Support of AI Feedback in Practice Exams Transforms Student Learning Behaviors | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Mak Ahmad, Prerna Ravi, David Karger, and Marc Facciotti. 2025. How Adding Metacognitive Requirements in Support of AI Feedback in Practice Exams Transforms Student Learning Behaviors. In Proceedings of the Twelfth ACM Conference on Learning @ Scale (L@S '25). Association for Computing Machinery, New York, NY, USA, 164–175. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science | en_US |
| dc.identifier.mitlicense | PUBLISHER_POLICY | |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
| eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
| dc.date.updated | 2025-08-01T08:01:11Z | |
| dc.language.rfc3066 | en | |
| dc.rights.holder | The author(s) | |
| dspace.date.submission | 2025-08-01T08:01:11Z | |
| mit.license | PUBLISHER_CC | |
| mit.metadata.status | Authority Work and Publication Information Needed | en_US |