The Data Efficiency of Deep Learning Is Degraded by Unnecessary Input Dimensions
Author(s)
D'Amario, Vanessa; Srivastava, Sanjana; Sasaki, Tomotake; Boix Bosch, Xavier
Downloadfncom-16-760085.pdf (1.727Mb)
Publisher with Creative Commons License
Publisher with Creative Commons License
Creative Commons Attribution
Terms of use
Metadata
Show full item recordAbstract
Biological learning systems are outstanding in their ability to learn from limited training data compared to the most successful learning machines, <jats:italic>i.e.</jats:italic>, Deep Neural Networks (DNNs). What are the key aspects that underlie this data efficiency gap is an unresolved question at the core of biological and artificial intelligence. We hypothesize that one important aspect is that biological systems rely on mechanisms such as foveations in order to reduce unnecessary input dimensions for the task at hand, <jats:italic>e.g.</jats:italic>, background in object recognition, while state-of-the-art DNNs do not. Datasets to train DNNs often contain such unnecessary input dimensions, and these lead to more trainable parameters. Yet, it is not clear whether this affects the DNNs' data efficiency because DNNs are robust to increasing the number of parameters in the hidden layers, and it is uncertain whether this holds true for the input layer. In this paper, we investigate the impact of unnecessary input dimensions on the DNNs data efficiency, namely, the amount of examples needed to achieve certain generalization performance. Our results show that unnecessary input dimensions that are task-unrelated substantially degrade data efficiency. This highlights the need for mechanisms that remove task-unrelated dimensions, such as foveation for image classification, in order to enable data efficiency gains.
Date issued
2022-01Department
Massachusetts Institute of Technology. Department of Brain and Cognitive SciencesJournal
Frontiers in Computational Neuroscience
Publisher
Frontiers Media SA
Citation
D'Amario, Vanessa, Srivastava, Sanjana, Sasaki, Tomotake and Boix, Xavier. 2022. "The Data Efficiency of Deep Learning Is Degraded by Unnecessary Input Dimensions." Frontiers in Computational Neuroscience, 16.
Version: Final published version
ISSN
1662-5188