Show simple item record

dc.contributor.authorDeStefano, Timothy
dc.contributor.authorKellogg, Katherine C.
dc.contributor.authorMenietti, Michael
dc.contributor.authorVendraminelli, Luca
dc.date.accessioned2022-10-12T21:07:46Z
dc.date.available2022-10-12T21:07:46Z
dc.date.issued2022-10-12
dc.identifier.urihttps://hdl.handle.net/1721.1/145813
dc.description.abstractHow is algorithmic model interpretability related to human acceptance of algorithmic recommendations and performance on decision-making tasks? We explored these questions in a multi-method field study of a large multinational fashion organization. We first conducted a quantitative field experiment to compare the use of two models—an interpretable versus an uninterpretable algorithmic model— designed to assist employees with decision making around how many products to send to each of its stores. Contrary to what the literature on interpretable algorithms would lead us to expect, under conditions of high perceived uncertainty, decision makers’ use of an uninterpretable algorithmic model was associated with higher acceptance of algorithmic recommendations and higher task performance than was their use of an interpretable algorithmic model with a similar level of performance. We next investigated this puzzling result using 31 interviews with 14 employees—2 algorithm developers, 2 managers, and 10 decision makers. We advance two concepts that suggest a refinement of theory on interpretable algorithms. First, overconfident troubleshooting—a decision maker rejecting a recommendation coming from an interpretable algorithm, because of their belief that they understand the inner workings of complex processes better than they actually do. Second, social proofing the algorithm—including respected peers in the algorithm development and testing process—may make it more likely that decision makers accept recommendations coming from an uninterpretable algorithm in situations characterized by high perceived uncertainty, because the decision makers may seek to reduce their uncertainty by incorporating the opinions of people with their own knowledge base and experience.en_US
dc.language.isoen_USen_US
dc.relation.ispartofseriesMIT Sloan School of Management Working Paper;6797-22
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 United States*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/us/*
dc.subjectInterpretable AI, Artificial Intelligence, Machine Learning, Algorithm Aversion, AI Adoption, Firm Productivity, AI and Strategy, Human-in-the-loop Decision Makingen_US
dc.titleWhy Providing Humans with Interpretable Algorithms May, Counterintuitively, Lead to Lower Decision-making Performanceen_US
dc.typeWorking Paperen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record