Show simple item record

dc.contributor.authorLiao, Qianli
dc.contributor.authorMiranda, Brando
dc.contributor.authorHidary, Jack
dc.contributor.authorPoggio, Tomaso
dc.date.accessioned2018-07-11T18:15:36Z
dc.date.available2018-07-11T18:15:36Z
dc.date.issued2018-07-11
dc.identifier.urihttp://hdl.handle.net/1721.1/116911
dc.description.abstractDeep networks are usually trained and tested in a regime in which the training classification error is not a good predictor of the test error. Thus the consensus has been that generalization, defined as convergence of the empirical to the expected error, does not hold for deep networks. Here we show that, when normalized appropriately after training, deep networks trained on exponential type losses show a good linear dependence of test loss on training loss. The observation, motivated by a previous theoretical analysis of overparameterization and overfitting, not only demonstrates the validity of classical generalization bounds for deep learning but suggests that they are tight. In addition, we also show that the bound of the classification error by the normalized cross entropy loss is empirically rather tight on the data sets we studied.en_US
dc.description.sponsorshipThis material is based upon work supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216.en_US
dc.language.isoen_USen_US
dc.publisherCenter for Brains, Minds and Machines (CBMM)en_US
dc.relation.ispartofseriesCBMM Memo Series;091
dc.titleClassical generalization bounds are surprisingly tight for Deep Networksen_US
dc.typeTechnical Reporten_US
dc.typeWorking Paperen_US
dc.typeOtheren_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record