| dc.contributor.author | Schrimpf, Martin | |
| dc.contributor.author | Blank, Idan Asher | |
| dc.contributor.author | Tuckute, Greta | |
| dc.contributor.author | Kauf, Carina | |
| dc.contributor.author | Hosseini, Eghbal A | |
| dc.contributor.author | Kanwisher, Nancy | |
| dc.contributor.author | Tenenbaum, Joshua B | |
| dc.contributor.author | Fedorenko, Evelina | |
| dc.date.accessioned | 2021-11-23T17:26:58Z | |
| dc.date.available | 2021-11-23T17:26:58Z | |
| dc.date.issued | 2021-11-09 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/138214 | |
| dc.description.abstract | <jats:p>The neuroscience of perception has recently been revolutionized with an integrative modeling approach in which computation, brain function, and behavior are linked across many datasets and many computational models. By revealing trends across models, this approach yields novel insights into cognitive and neural mechanisms in the target domain. We here present a systematic study taking this approach to higher-level cognition: human language processing, our species’ signature cognitive skill. We find that the most powerful “transformer” models predict nearly 100% of explainable variance in neural responses to sentences and generalize across different datasets and imaging modalities (functional MRI and electrocorticography). Models’ neural fits (“brain score”) and fits to behavioral responses are both strongly correlated with model accuracy on the next-word prediction task (but not other language tasks). Model architecture appears to substantially contribute to neural fit. These results provide computationally explicit evidence that predictive processing fundamentally shapes the language comprehension mechanisms in the human brain.</jats:p> | en_US |
| dc.language.iso | en | |
| dc.publisher | Proceedings of the National Academy of Sciences | en_US |
| dc.relation.isversionof | 10.1073/pnas.2105646118 | en_US |
| dc.rights | Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. | en_US |
| dc.source | PNAS | en_US |
| dc.title | The neural architecture of language: Integrative modeling converges on predictive processing | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Schrimpf, Martin, Blank, Idan Asher, Tuckute, Greta, Kauf, Carina, Hosseini, Eghbal A et al. 2021. "The neural architecture of language: Integrative modeling converges on predictive processing." Proceedings of the National Academy of Sciences, 118 (45). | |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences | |
| dc.contributor.department | McGovern Institute for Brain Research at MIT | |
| dc.contributor.department | Center for Brains, Minds, and Machines | |
| dc.relation.journal | Proceedings of the National Academy of Sciences | en_US |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/JournalArticle | en_US |
| eprint.status | http://purl.org/eprint/status/PeerReviewed | en_US |
| dc.date.updated | 2021-11-23T17:22:16Z | |
| dspace.orderedauthors | Schrimpf, M; Blank, IA; Tuckute, G; Kauf, C; Hosseini, EA; Kanwisher, N; Tenenbaum, JB; Fedorenko, E | en_US |
| dspace.date.submission | 2021-11-23T17:22:18Z | |
| mit.journal.volume | 118 | en_US |
| mit.journal.issue | 45 | en_US |
| mit.license | PUBLISHER_POLICY | |
| mit.metadata.status | Authority Work and Publication Information Needed | en_US |