Show simple item record

dc.contributor.authorBao, Yujia
dc.contributor.authorChang, Shiyu
dc.contributor.authorYu, Mo
dc.contributor.authorBarzilay, Regina
dc.date.accessioned2021-02-09T22:36:48Z
dc.date.available2021-02-09T22:36:48Z
dc.date.issued2018-10
dc.identifier.urihttps://hdl.handle.net/1721.1/129732
dc.description.abstractAttention-based models are successful when trained on large amounts of data. In this paper, we demonstrate that even in the low-resource scenario, attention can be learned effectively. To this end, we start with discrete human-annotated rationales and map them into continuous attention. Our central hypothesis is that this mapping is general across domains, and thus can be transferred from resource-rich domains to low-resource ones. Our model jointly learns a domain-invariant representation and induces the desired mapping between rationales and attention. Our empirical results validate this hypothesis and show that our approach delivers significant gains over state-of-the-art baselines, yielding over 15% average error reduction on benchmark datasets.en_US
dc.language.isoen
dc.publisherAssociation for Computational Linguistics (ACL)en_US
dc.relation.isversionofhttp://dx.doi.org/10.18653/v1/d18-1216en_US
dc.rightsCreative Commons Attribution 4.0 International licenseen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceAssociation for Computational Linguisticsen_US
dc.titleDeriving Machine Attention from Human Rationalesen_US
dc.typeArticleen_US
dc.identifier.citationBao, Yujia et al. "Deriving Machine Attention from Human Rationales." Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, October-November 2018, Brussels, Belgium, Association for Computational Linguistics, October 2018. © 2018 Association for Computational Linguisticsen_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.relation.journalProceedings of the 2018 Conference on Empirical Methods in Natural Language Processingen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2020-12-01T12:55:59Z
dspace.orderedauthorsBao, Y; Chang, S; Yu, M; Barzilay, Ren_US
dspace.date.submission2020-12-01T12:56:04Z
mit.licensePUBLISHER_CC
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record