Notice

This is not the latest version of this item. The latest version can be found at:https://dspace.mit.edu/handle/1721.1/139838.3

Show simple item record

dc.contributor.authorD'Amario, Vanessa
dc.contributor.authorSrivastava, Sanjana
dc.contributor.authorSasaki, Tomotake
dc.contributor.authorBoix, Xavier
dc.date.accessioned2022-02-02T17:27:58Z
dc.date.available2022-02-02T17:27:58Z
dc.date.issued2022-01-31
dc.identifier.issn1662-5188
dc.identifier.urihttps://hdl.handle.net/1721.1/139838
dc.description.abstractBiological learning systems are outstanding in their ability to learn from limited training data compared to the most successful learning machines, <jats:italic>i.e.</jats:italic>, Deep Neural Networks (DNNs). What are the key aspects that underlie this data efficiency gap is an unresolved question at the core of biological and artificial intelligence. We hypothesize that one important aspect is that biological systems rely on mechanisms such as foveations in order to reduce unnecessary input dimensions for the task at hand, <jats:italic>e.g.</jats:italic>, background in object recognition, while state-of-the-art DNNs do not. Datasets to train DNNs often contain such unnecessary input dimensions, and these lead to more trainable parameters. Yet, it is not clear whether this affects the DNNs' data efficiency because DNNs are robust to increasing the number of parameters in the hidden layers, and it is uncertain whether this holds true for the input layer. In this paper, we investigate the impact of unnecessary input dimensions on the DNNs data efficiency, namely, the amount of examples needed to achieve certain generalization performance. Our results show that unnecessary input dimensions that are task-unrelated substantially degrade data efficiency. This highlights the need for mechanisms that remove task-unrelated dimensions, such as foveation for image classification, in order to enable data efficiency gains.en_US
dc.publisherFrontiers Media SAen_US
dc.relation.isversionof10.3389/fncom.2022.760085en_US
dc.rightsCreative Commons Attribution 4.0 International licenseen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceFrontiersen_US
dc.titleThe Data Efficiency of Deep Learning Is Degraded by Unnecessary Input Dimensionsen_US
dc.typeArticleen_US
dc.identifier.citationD'Amario, Vanessa, Srivastava, Sanjana, Sasaki, Tomotake and Boix, Xavier. 2022. "The Data Efficiency of Deep Learning Is Degraded by Unnecessary Input Dimensions." Frontiers in Computational Neuroscience, 16.
dc.relation.journalFrontiers in Computational Neuroscienceen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dspace.date.submission2022-02-02T17:21:55Z
mit.journal.volume16en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

VersionItemDateSummary

*Selected version