Show simple item record

dc.contributor.authorSchoenegger, Philipp
dc.contributor.authorPark, Peter
dc.contributor.authorKarger, Ezra
dc.contributor.authorTrott, Sean
dc.contributor.authorTetlock, Philip
dc.date.accessioned2025-01-24T20:49:23Z
dc.date.available2025-01-24T20:49:23Z
dc.date.submitted2024-12-13
dc.identifier.issn2160-6455
dc.identifier.urihttps://hdl.handle.net/1721.1/158063
dc.description.abstractLarge language models (LLMs) match and sometimes exceed human performance in many domains. This study explores the potential of LLMs to augment human judgment in a forecasting task. We evaluate the effect on human forecasters of two LLM assistants: one designed to provide high-quality ("superforecasting") advice, and the other designed to be overconfident and base-rate neglecting, thus providing noisy forecasting advice. We compare participants using these assistants to a control group that received a less advanced model that did not provide numerical predictions or engage in explicit discussion of predictions. Participants (N = 991) answered a set of six forecasting questions and had the option to consult their assigned LLM assistant throughout. Our preregistered analyses show that interacting with each of our frontier LLM assistants significantly enhances prediction accuracy by between 24% and 28% compared to the control group. Exploratory analyses showed a pronounced outlier effect in one forecasting item, without which we find that the superforecasting assistant increased accuracy by 41%, compared with 29% for the noisy assistant. We further examine whether LLM forecasting augmentation disproportionately benefits less skilled forecasters, degrades the wisdom-of-the-crowd by reducing prediction diversity, or varies in effectiveness with question difficulty. Our data do not consistently support these hypotheses. Our results suggest that access to a frontier LLM assistant, even a noisy one, can be a helpful decision aid in cognitively demanding tasks compared to a less powerful model that does not provide specific forecasting advice. However, the effects of outliers suggest that further research into the robustness of this pattern is needed.en_US
dc.publisherACMen_US
dc.relation.isversionofhttp://dx.doi.org/10.1145/3707649en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleAI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracyen_US
dc.typeArticleen_US
dc.identifier.citationSchoenegger, Philipp, Park, Peter, Karger, Ezra, Trott, Sean and Tetlock, Philip. "AI-Augmented Predictions: LLM Assistants Improve Human Forecasting Accuracy." ACM Transactions on Interactive Intelligent Systems.
dc.relation.journalACM Transactions on Interactive Intelligent Systemsen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-01-01T08:46:22Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-01-01T08:46:22Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record