| dc.contributor.author | Eisape, Tiwalayo | |
| dc.contributor.author | Zaslavsky, Noga | |
| dc.contributor.author | Levy, Roger | |
| dc.date.accessioned | 2022-02-03T16:33:14Z | |
| dc.date.available | 2021-12-01T17:36:44Z | |
| dc.date.available | 2022-02-03T16:33:14Z | |
| dc.date.issued | 2020-11 | |
| dc.identifier.uri | https://hdl.handle.net/1721.1/138277.2 | |
| dc.description.abstract | Contemporary autoregressive language models (LMs) trained purely on corpus data have
been shown to capture numerous features of
human incremental processing. However, past
work has also suggested dissociations between
corpus probabilities and human next-word predictions. Here we evaluate several state-of-theart language models for their match to human
next-word predictions and to reading time behavior from eye movements. We then propose
a novel method for distilling the linguistic information implicit in human linguistic predictions into pre-trained LMs: Cloze Distillation.
We apply this method to a baseline neural LM
and show potential improvement in reading
time prediction and generalization to held-out
human cloze data. | en_US |
| dc.language.iso | en | |
| dc.publisher | Association for Computational Linguistics (ACL) | en_US |
| dc.relation.isversionof | http://dx.doi.org/10.18653/V1/2020.CONLL-1.49 | en_US |
| dc.rights | Creative Commons Attribution 4.0 International license | en_US |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | en_US |
| dc.source | Association for Computational Linguistics | en_US |
| dc.title | Cloze Distillation: Improving Neural Language Models with Human Next-Word Prediction | en_US |
| dc.type | Article | en_US |
| dc.identifier.citation | Eisape, Tiwalayo, Zaslavsky, Noga and Levy, Roger. 2020. "Cloze Distillation: Improving Neural Language Models with Human Next-Word Prediction." Proceedings of the 24th Conference on Computational Natural Language Learning. | en_US |
| dc.contributor.department | Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences | |
| dc.contributor.department | Center for Brains, Minds, and Machines | |
| dc.relation.journal | Proceedings of the 24th Conference on Computational Natural Language Learning | en_US |
| dc.eprint.version | Final published version | en_US |
| dc.type.uri | http://purl.org/eprint/type/ConferencePaper | en_US |
| eprint.status | http://purl.org/eprint/status/NonPeerReviewed | en_US |
| dc.date.updated | 2021-12-01T17:32:40Z | |
| dspace.orderedauthors | Eisape, T; Zaslavsky, N; Levy, R | en_US |
| dspace.date.submission | 2021-12-01T17:32:41Z | |
| mit.license | PUBLISHER_CC | |
| mit.metadata.status | Authority Work Needed | en_US |