| dc.contributor.author | Danry, Valdemar | |
| dc.contributor.author | Pataranutaporn, Pat | |
| dc.contributor.author | Groh, Matthew | |
| dc.contributor.author | Epstein, Ziv | |
| dc.date.accessioned | 2025-09-22T18:02:08Z | |
| dc.date.available | 2025-09-22T18:02:08Z | |
| dc.date.issued | 2025-04-25 | |
| dc.identifier.isbn | 979-8-4007-1394-1 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/162775 | |
| dc.description | CHI ’25, Yokohama, Japan | en_US |
| dc.description.abstract | Advanced Artificial Intelligence (AI) systems, specifically large language models (LLMs), have the capability to generate not just misinformation, but also deceptive explanations that can justify and propagate false information and discredit true information. We examined the impact of deceptive AI generated explanations on individuals’ beliefs in a pre-registered online experiment with 11,780 observations from 589 participants. We found that in addition to being more persuasive than accurate and honest explanations, AI-generated deceptive explanations can significantly amplify belief in false news headlines and undermine true ones as compared to AI systems that simply classify the headline incorrectly as being true/false. Moreover, our results show that logically invalid explanations are deemed less credible - diminishing the effects of deception. This underscores the importance of teaching logical reasoning and critical thinking skills to identify logically invalid arguments, fostering greater resilience against advanced AI-driven misinformation. | en_US |
| dc.publisher | ACM|CHI Conference on Human Factors in Computing Systems | en_US |
| dc.relation.isversionof | https://doi.org/10.1145/3706598.3713408 | en_US |
| dc.rights | Creative Commons Attribution | en_US |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | en_US |
| dc.source | Association for Computing Machinery | en_US |
| dc.title | Deceptive Explanations by Large Language Models Lead People to Change their Beliefs About Misinformation More Often than Honest Explanations | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Valdemar Danry, Pat Pataranutaporn, Matthew Groh, and Ziv Epstein. 2025. Deceptive Explanations by Large Language Models Lead People to Change their Beliefs About Misinformation More Often than Honest Explanations. In Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems (CHI '25). Association for Computing Machinery, New York, NY, USA, Article 933, 1–31. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Media Laboratory | en_US |
| dc.identifier.mitlicense | PUBLISHER_POLICY | |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
| eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
| dc.date.updated | 2025-08-01T08:08:29Z | |
| dc.language.rfc3066 | en | |
| dc.rights.holder | The author(s) | |
| dspace.date.submission | 2025-08-01T08:08:30Z | |
| mit.license | PUBLISHER_CC | |
| mit.metadata.status | Authority Work and Publication Information Needed | en_US |