Show simple item record

dc.contributor.authorZhang, Yue
dc.contributor.authorWeninger, Felix
dc.contributor.authorBjorn, Schuller
dc.contributor.authorPicard, Rosalind W.
dc.date.accessioned2020-02-13T20:24:53Z
dc.date.available2020-02-13T20:24:53Z
dc.date.issued2019-12
dc.identifier.issn1949-3045
dc.identifier.issn2371-9850
dc.identifier.urihttps://hdl.handle.net/1721.1/123806
dc.description.abstractHumans perceive emotion from each other using a holistic perspective, accounting for diverse personal, non-emotional variables that shape expression. In contrast, today's algorithms are mainly designed to recognize emotion in isolation. In this work, we propose a multi-task learning approach to jointly learn the recognition of affective states from speech along with various speaker attributes. A problem with multi-task learning is that sometimes inductive transfer can negatively impact performance. To mitigate negative transfer, we introduce the Paralinguistic Non-metric Dimensional Analysis (PaNDA) method that systematically measures task relatedness and also enables visualizing the topology of affective phenomena as a whole. In addition, we present a generic framework that conflates the concepts of single-task and multi-task learning. Using this framework, we construct two models that demonstrate holistic affect recognition: one treats all tasks as equally related, whereas the other one incorporates the task correlations between a main task and its supporting tasks obtained from PaNDA. Both models employ a multi-task deep neural network, in which separate output layers are used to predict discrete and continuous attributes, while hidden layers are shared across different tasks. On average across 18 classification and regression tasks, the weighted multi-task learning with PaNDA significantly improves performance compared to single-task and unweighted multi-task learning.en_US
dc.description.sponsorshipEUen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.isversionofhttp://dx.doi.org/10.1109/taffc.2019.2961881en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceProf. Picard via Elizabeth Soergelen_US
dc.titleHolistic Affect Recognition Using PaNDA: Paralinguistic Non-metric Dimensional Analysisen_US
dc.typeArticleen_US
dc.identifier.citationZhang, Y. et al. "Holistic Affect Recognition Using PaNDA: Paralinguistic Non-metric Dimensional Analysis," IEEE Transactions on Affective Computing (December 2019). © 2019 IEEEen_US
dc.contributor.departmentMassachusetts Institute of Technology. Media Laboratoryen_US
dc.relation.journalIEEE Transactions on Affective Computingen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dspace.date.submission2020-02-08T15:03:27Z
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusComplete


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record