Show simple item record

dc.contributor.authorCano-Córdoba, Felipe
dc.contributor.authorSarma, Sanjay
dc.contributor.authorSubirana, Brian
dc.date.accessioned2018-02-12T21:29:47Z
dc.date.available2018-02-12T21:29:47Z
dc.date.issued2017-12-05
dc.identifier.urihttp://hdl.handle.net/1721.1/113608
dc.description.abstractIn [42] we suggested that any memory stored in the human/animal brain is forgotten following the Ebingghaus curve – in this follow-on paper, we define a novel algebraic structure, a Forgetting Neural Network, as a simple mathematical model based on assuming parameters of a neuron in a neural network are forgotten using the Ebbinghaus forgetting curve. We model neural networks in Sobolev spaces using [35] as our departure point and demonstrate four novel theorems of Forgetting Neural Networks: theorem of non-instantaneous forgetting, theorem of universal forgetting, curse of forgetting theorem, and center of mass theorem. We also proof the novel decreasing inference theorem which we feel is relevant beyond Ebbinghaus forgetting: compositional deep neural networks cannot arbitrarily combine low level “features” – meaning only certain arrangements of features calculated in intermediate levels can show up in higher levels. This proof leads us to present the possibly most efficient representation of neural networks’ “minimal polynomial basis layer” (MPBL) since our basis construct can generate n polynomials of order m using only 2m + 1 + n neurons. As we briefly discuss in the conclusion, there are about 10 similarities between forgetting neural networks and human forgetting and our research elicits more questions than it answers and may have implications for neuroscience research including our understanding of how babies learn (or, perhaps, forget), including what we call the baby forgetting conjecture.en_US
dc.description.sponsorshipThis work was supported by the Center for Brains, Minds and Ma- chines (CBMM), funded by NSF STC award CCF - 1231216.en_US
dc.language.isoen_USen_US
dc.publisherCenter for Brains, Minds and Machines (CBMM)en_US
dc.relation.ispartofseriesCBMM Memo Series;071
dc.rightsAttribution-NonCommercial-ShareAlike 3.0 United States*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/3.0/us/*
dc.subjectForgetting Neural Networken_US
dc.subjectdecreasing inference theoremen_US
dc.subjecttheorem of non-instantaneous forgettingen_US
dc.subjecttheorem of universal forgettingen_US
dc.subjectcurse of forgetting theoremen_US
dc.subjectcenter of mass theoremen_US
dc.titleTheory of Intelligence with Forgetting: Mathematical Theorems Explaining Human Universal Forgetting using “Forgetting Neural Networks”en_US
dc.typeTechnical Reporten_US
dc.typeWorking Paperen_US
dc.typeOtheren_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record