Notice

This is not the latest version of this item. The latest version can be found at:https://dspace.mit.edu/handle/1721.1/137379.2

Show simple item record

dc.contributor.authorLee, JKW
dc.contributor.authorSattigeri, P
dc.contributor.authorWornell, GW
dc.date.accessioned2021-11-04T17:21:35Z
dc.date.available2021-11-04T17:21:35Z
dc.date.issued2019-12
dc.identifier.urihttps://hdl.handle.net/1721.1/137379
dc.description.abstract© 2019 Neural information processing systems foundation. All rights reserved. The advent of deep learning algorithms for mobile devices and sensors has led to a dramatic expansion in the availability and number of systems trained on a wide range of machine learning tasks, creating a host of opportunities and challenges in the realm of transfer learning. Currently, most transfer learning methods require some kind of control over the systems learned, either by enforcing constraints during the source training, or through the use of a joint optimization objective between tasks that requires all data be co-located for training. However, for practical, privacy, or other reasons, in a variety of applications we may have no control over the individual source task training, nor access to source training samples. Instead we only have access to features pre-trained on such data as the output of “black-boxes.” For such scenarios, we consider the multi-source learning problem of training a classifier using an ensemble of pre-trained neural networks for a set of classes that have not been observed by any of the source networks, and for which we have very few training samples. We show that by using these distributed networks as feature extractors, we can train an effective classifier in a computationally-efficient manner using tools from (nonlinear) maximal correlation analysis. In particular, we develop a method we refer to as maximal correlation weighting (MCW) to build the required target classifier from an appropriate weighting of the feature functions from the source networks. We illustrate the effectiveness of the resulting classifier on datasets derived from the CIFAR-100, Stanford Dogs, and Tiny ImageNet datasets, and, in addition, use the methodology to characterize the relative value of different source tasks in learning a target task.en_US
dc.language.isoen
dc.relation.isversionofhttps://papers.nips.cc/paper/2019/hash/6048ff4e8cb07aa60b6777b6f7384d52-Abstract.htmlen_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceNeural Information Processing Systems (NIPS)en_US
dc.titleLearning new tricks from old dogs: Multi-source transfer learning from pre-trained networksen_US
dc.typeArticleen_US
dc.identifier.citationLee, JKW, Sattigeri, P and Wornell, GW. 2019. "Learning new tricks from old dogs: Multi-source transfer learning from pre-trained networks." Advances in Neural Information Processing Systems, 32.
dc.relation.journalAdvances in Neural Information Processing Systemsen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2021-02-03T17:11:48Z
dspace.orderedauthorsLee, JKW; Sattigeri, P; Wornell, GWen_US
dspace.date.submission2021-02-03T17:11:50Z
mit.journal.volume32en_US
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

VersionItemDateSummary

*Selected version