Show simple item record

dc.contributor.authorMhaskar, Hrushikesh
dc.contributor.authorPoggio, Tomaso
dc.date.accessioned2016-08-12T22:44:41Z
dc.date.available2016-08-12T22:44:41Z
dc.date.issued2016-08-12
dc.identifier.urihttp://hdl.handle.net/1721.1/103911
dc.description.abstractThe paper briefly reviews several recent results on hierarchical architectures for learning from examples, that may formally explain the conditions under which Deep Convolutional Neural Networks perform much better in function approximation problems than shallow, one-hidden layer architectures. The paper announces new results for a non-smooth activation function – the ReLU function – used in present-day neural networks, as well as for the Gaussian networks. We propose a new definition of relative dimension to encapsulate different notions of sparsity of a function class that can possibly be exploited by deep networks but not by shallow ones to drastically reduce the complexity required for approximation and learning.en_US
dc.description.sponsorshipThis work was supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF – 1231216.en_US
dc.language.isoen_USen_US
dc.publisherCenter for Brains, Minds and Machines (CBMM), arXiven_US
dc.relation.ispartofseriesCBMM Memo Series;054
dc.rightsAttribution-NonCommercial-ShareAlike 3.0 United States*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/3.0/us/*
dc.subjecthierarchical architecturesen_US
dc.subjectDeep Convolutional Neural Networksen_US
dc.subjectReLU functionen_US
dc.subjectGaussian networksen_US
dc.titleDeep vs. shallow networks : An approximation theory perspectiveen_US
dc.typeTechnical Reporten_US
dc.typeWorking Paperen_US
dc.typeOtheren_US
dc.identifier.citationarXiv:1608.03287en_US
dc.audience.educationlevel


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record