Show simple item record

dc.contributor.authorRecasens, Adrià
dc.contributor.authorKellnhofer, Petr
dc.contributor.authorStent, Simon
dc.contributor.authorMatusik, Wojciech
dc.contributor.authorTorralba, Antonio
dc.date.accessioned2021-11-09T12:18:19Z
dc.date.available2021-11-09T12:18:19Z
dc.date.issued2018
dc.identifier.issn0302-9743
dc.identifier.issn1611-3349
dc.identifier.urihttps://hdl.handle.net/1721.1/137841
dc.description.abstract© Springer Nature Switzerland AG 2018. We introduce a saliency-based distortion layer for convolutional neural networks that helps to improve the spatial sampling of input data for a given task. Our differentiable layer can be added as a preprocessing block to existing task networks and trained altogether in an end-to-end fashion. The effect of the layer is to efficiently estimate how to sample from the original data in order to boost task performance. For example, for an image classification task in which the original data might range in size up to several megapixels, but where the desired input images to the task network are much smaller, our layer learns how best to sample from the underlying high resolution data in a manner which preserves task-relevant information better than uniform downsampling. This has the effect of creating distorted, caricature-like intermediate images, in which idiosyncratic elements of the image that improve task performance are zoomed and exaggerated. Unlike alternative approaches such as spatial transformer networks, our proposed layer is inspired by image saliency, computed efficiently from uniformly downsampled data, and degrades gracefully to a uniform sampling strategy under uncertainty. We apply our layer to improve existing networks for the tasks of human gaze estimation and fine-grained object classification. Code for our method is available in: http://github.com/recasens/Saliency-Sampler.en_US
dc.language.isoen
dc.publisherSpringer International Publishingen_US
dc.relation.isversionof10.1007/978-3-030-01240-3_4en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceComputer Vision Foundationen_US
dc.titleLearning to Zoom: a Saliency-Based Sampling Layer for Neural Networksen_US
dc.typeArticleen_US
dc.identifier.citationRecasens, Adrià, Kellnhofer, Petr, Stent, Simon, Matusik, Wojciech and Torralba, Antonio. 2018. "Learning to Zoom: a Saliency-Based Sampling Layer for Neural Networks."
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2019-06-21T16:26:05Z
dspace.date.submission2019-06-21T16:26:07Z
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record