Why Providing Humans with Interpretable Algorithms May, Counterintuitively, Lead to Lower Decision-making Performance
Author(s)
DeStefano, Timothy; Kellogg, Katherine C.; Menietti, Michael; Vendraminelli, Luca
DownloadDKMV_WP_2022_Providing_Humans_Interpretable_Algorithms.pdf (753.5Kb)
Terms of use
Metadata
Show full item recordAbstract
How is algorithmic model interpretability related to human acceptance of algorithmic recommendations and performance on decision-making tasks? We explored these questions in a multi-method field study of a large multinational fashion organization. We first conducted a quantitative field experiment to compare the use of two models—an interpretable versus an uninterpretable algorithmic model— designed to assist employees with decision making around how many products to send to each of its stores. Contrary to what the literature on interpretable algorithms would lead us to expect, under conditions of high perceived uncertainty, decision makers’ use of an uninterpretable algorithmic model was associated with higher acceptance of algorithmic recommendations and higher task performance than was their use of an interpretable algorithmic model with a similar level of performance. We next investigated this puzzling result using 31 interviews with 14 employees—2 algorithm developers, 2 managers, and 10 decision makers. We advance two concepts that suggest a refinement of theory on interpretable algorithms. First, overconfident troubleshooting—a decision maker rejecting a recommendation coming from an interpretable algorithm, because of their belief that they understand the inner workings of complex processes better than they actually do. Second, social proofing the algorithm—including respected peers in the algorithm development and testing process—may make it more likely that decision makers accept recommendations coming from an uninterpretable algorithm in situations characterized by high perceived uncertainty, because the decision makers may seek to reduce their uncertainty by incorporating the opinions of people with their own knowledge base and experience.
Date issued
2022-10-12Series/Report no.
MIT Sloan School of Management Working Paper;6797-22
Keywords
Interpretable AI, Artificial Intelligence, Machine Learning, Algorithm Aversion, AI Adoption, Firm Productivity, AI and Strategy, Human-in-the-loop Decision Making
Collections
The following license files are associated with this item: