Show simple item record

dc.contributor.authorWang, Lichen
dc.contributor.authorWu, Jiaxiang
dc.contributor.authorHuang, Shao-Lun
dc.contributor.authorZheng, Lizhong
dc.contributor.authorXu, Xiangxiang
dc.contributor.authorZhang, Lin
dc.contributor.authorHuang, Junzhou
dc.date.accessioned2021-11-08T19:30:22Z
dc.date.available2021-11-08T19:30:22Z
dc.date.issued2019
dc.identifier.urihttps://hdl.handle.net/1721.1/137795
dc.description.abstractCopyright © 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. One primary focus in multimodal feature extraction is to find the representations of individual modalities that are maximally correlated. As a well-known measure of dependence, the Hirschfeld-Gebelein-Rényi (HGR) maximal correlation becomes an appealing objective because of its operational meaning and desirable properties. However, the strict whitening constraints formalized in the HGR maximal correlation limit its application. To address this problem, this paper proposes Soft-HGR, a novel framework to extract informative features from multiple data modalities. Specifically, our framework prevents the “hard” whitening constraints, while simultaneously preserving the same feature geometry as in the HGR maximal correlation. The objective of Soft-HGR is straightforward, only involving two inner products, which guarantees the efficiency and stability in optimization. We further generalize the framework to handle more than two modalities and missing modalities. When labels are partially available, we enhance the discriminative power of the feature representations by making a semi-supervised adaptation. Empirical evaluation implies that our approach learns more informative feature mappings and is more efficient to optimize.en_US
dc.language.isoen
dc.publisherAssociation for the Advancement of Artificial Intelligence (AAAI)en_US
dc.relation.isversionof10.1609/AAAI.V33I01.33015281en_US
dc.rightsCreative Commons Attribution-Noncommercial-Share Alikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourcearXiven_US
dc.titleAn Efficient Approach to Informative Feature Extraction from Multimodal Dataen_US
dc.typeArticleen_US
dc.identifier.citationWang, Lichen, Wu, Jiaxiang, Huang, Shao-Lun, Zheng, Lizhong, Xu, Xiangxiang et al. 2019. "An Efficient Approach to Informative Feature Extraction from Multimodal Data." Proceedings of the AAAI Conference on Artificial Intelligence, 33.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.relation.journalProceedings of the AAAI Conference on Artificial Intelligenceen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-01-25T19:11:05Z
dspace.orderedauthorsWang, L; Wu, J; Huang, S-L; Zheng, L; Xu, X; Zhang, L; Huang, Jen_US
dspace.date.submission2021-01-25T19:11:13Z
mit.journal.volume33en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record