Show simple item record

dc.contributor.authorShen, Jocelyn
dc.contributor.authorDiPaola, Daniella
dc.contributor.authorAli, Safinah
dc.contributor.authorSap, Maarten
dc.contributor.authorPark, Hae Won
dc.contributor.authorBreazeal, Cynthia
dc.date.accessioned2025-01-28T21:45:28Z
dc.date.available2025-01-28T21:45:28Z
dc.date.issued2024
dc.identifier.urihttps://hdl.handle.net/1721.1/158124
dc.description.abstractBackground: Empathy is a driving force in our connection to others, our mental well-being, and resilience to challenges. With the rise of generative artificial intelligence (AI) systems, mental health chatbots, and AI social support companions, it is important to understand how empathy unfolds toward stories from human versus AI narrators and how transparency plays a role in user emotions. Objective: We aim to understand how empathy shifts across human-written versus AI-written stories, and how these findings inform ethical implications and human-centered design of using mental health chatbots as objects of empathy. Methods: We conducted crowd-sourced studies with 985 participants who each wrote a personal story and then rated empathy toward 2 retrieved stories, where one was written by a language model, and another was written by a human. Our studies varied disclosing whether a story was written by a human or an AI system to see how transparent author information affects empathy toward the narrator. We conducted mixed methods analyses: through statistical tests, we compared user’s self-reported state empathy toward the stories across different conditions. In addition, we qualitatively coded open-ended feedback about reactions to the stories to understand how and why transparency affects empathy toward human versus AI storytellers. Results: We found that participants significantly empathized with human-written over AI-written stories in almost all conditions, regardless of whether they are aware (t196=7.07, P<.001, Cohen d=0.60) or not aware (t298=3.46, P<.001, Cohen d=0.24) that an AI system wrote the story. We also found that participants reported greater willingness to empathize with AI-written stories when there was transparency about the story author (t494=–5.49, P<.001, Cohen d=0.36). Conclusions: Our work sheds light on how empathy toward AI or human narrators is tied to the way the text is presented, thus informing ethical considerations of empathetic artificial social support or mental health chatbots.en_US
dc.language.isoen
dc.publisherJMIR Publications Inc.en_US
dc.relation.isversionofhttps://doi.org/10.2196/62679en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceJMIR Publications Inc.en_US
dc.titleEmpathy Toward Artificial Intelligence Versus Human Experiences and the Role of Transparency in Mental Health and Social Support Chatbot Design: Comparative Studyen_US
dc.typeArticleen_US
dc.identifier.citationShen, Jocelyn, DiPaola, Daniella, Ali, Safinah, Sap, Maarten, Park, Hae Won et al. 2024. "Empathy Toward Artificial Intelligence Versus Human Experiences and the Role of Transparency in Mental Health and Social Support Chatbot Design: Comparative Study." JMIR Mental Health, 11.
dc.contributor.departmentMassachusetts Institute of Technology. Media Laboratoryen_US
dc.relation.journalJMIR Mental Healthen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-01-28T21:38:00Z
dspace.orderedauthorsShen, J; DiPaola, D; Ali, S; Sap, M; Park, HW; Breazeal, Cen_US
dspace.date.submission2025-01-28T21:38:02Z
mit.journal.volume11en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record