Size-independent sample complexity of neural networks
Author(s)
Golowich, Noah; Rakhlin, Alexander; Shamir, Ohad
DownloadAccepted version (310.5Kb)
Open Access Policy
Open Access Policy
Creative Commons Attribution-Noncommercial-Share Alike
Terms of use
Metadata
Show full item recordAbstract
We study the sample complexity of learning neural networks by providing new bounds on their Rademacher complexity, assuming norm constraints on the parameter matrix of each layer. Compared to previous work, these complexity bounds have improved dependence on the network depth and, under some additional assumptions, are fully independent of the network size (both depth and width). These results are derived using some novel techniques, which may be of independent interest.
Date issued
2020Department
Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences; Massachusetts Institute of Technology. Laboratory for Information and Decision Systems; Massachusetts Institute of Technology. Institute for Data, Systems, and Society; Statistics and Data Science Center (Massachusetts Institute of Technology)Journal
Information and Inference
Publisher
Oxford University Press (OUP)
Citation
Golowich, Noah, Rakhlin, Alexander and Shamir, Ohad. 2020. "Size-independent sample complexity of neural networks." Information and Inference, 9 (2).
Version: Author's final manuscript