Notice

This is not the latest version of this item. The latest version can be found at:https://dspace.mit.edu/handle/1721.1/136301.2

Show simple item record

dc.contributor.authorPoggio, Tomaso
dc.contributor.authorLiao, Qianli
dc.contributor.authorBanburski, Andrzej
dc.date.accessioned2021-10-27T20:34:47Z
dc.date.available2021-10-27T20:34:47Z
dc.date.issued2020
dc.identifier.urihttps://hdl.handle.net/1721.1/136301
dc.description.abstract© 2020, The Author(s). Overparametrized deep networks predict well, despite the lack of an explicit complexity control during training, such as an explicit regularization term. For exponential-type loss functions, we solve this puzzle by showing an effective regularization effect of gradient descent in terms of the normalized weights that are relevant for classification.
dc.language.isoen
dc.publisherSpringer Science and Business Media LLC
dc.relation.isversionof10.1038/S41467-020-14663-9
dc.rightsCreative Commons Attribution 4.0 International license
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.sourceNature
dc.titleComplexity control by gradient descent in deep networks
dc.typeArticle
dc.relation.journalNature Communications
dc.eprint.versionFinal published version
dc.type.urihttp://purl.org/eprint/type/JournalArticle
eprint.statushttp://purl.org/eprint/status/PeerReviewed
dc.date.updated2021-03-24T13:05:49Z
dspace.orderedauthorsPoggio, T; Liao, Q; Banburski, A
dspace.date.submission2021-03-24T13:05:50Z
mit.journal.volume11
mit.journal.issue1
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Needed


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

VersionItemDateSummary

*Selected version