Show simple item record

dc.contributor.authorMlynarski, Wiktor
dc.contributor.authorMcDermott, Joshua H.
dc.date.accessioned2018-04-03T14:49:06Z
dc.date.available2018-04-03T14:49:06Z
dc.date.issued2018-02
dc.identifier.issn0899-7667
dc.identifier.issn1530-888X
dc.identifier.urihttp://hdl.handle.net/1721.1/114502
dc.description.abstractInteraction with the world requires an organism to transform sensory signals into representations in which behaviorally meaningful properties of the environment are made explicit. These representations are derived through cascades of neuronal processing stages in which neurons at each stage recode the output of preceding stages. Explanations of sensory coding may thus involve understanding how low-level patterns are combined into more complex structures. To gain insight into such midlevel representations for sound, we designed a hierarchical generative model of natural sounds that learns combinations of spectrotemporal features from natural stimulus statistics. In the first layer, the model forms a sparse convolutional code of spectrograms using a dictionary of learned spectrotemporal kernels. To generalize from specific kernel activation patterns, the second layer encodes patterns of time-varying magnitude of multiple first-layer coefficients. When trained on corpora of speech and environmental sounds, some second-layer units learned to group similar spectrotemporal features. Others instantiate opponency between distinct sets of features. Such groupings might be instantiated by neurons in the auditory cortex, providing a hypothesis for midlevel neuronal computation.en_US
dc.description.sponsorshipNational Science Foundation (U.S.) (McGovern Institute for Brain Research at MIT. Center for Brains, Minds, and Machines. STC Award CCF-1231216)en_US
dc.description.sponsorshipJames S. McDonnell Foundation (Scholar Award)en_US
dc.publisherMIT Pressen_US
dc.relation.isversionofhttp://dx.doi.org/10.1162/neco_a_01048en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceMassachusetts Institute of Technology Pressen_US
dc.titleLearning Midlevel Auditory Codes from Natural Sound Statisticsen_US
dc.typeArticleen_US
dc.identifier.citationMłynarski, Wiktor, and Josh H. McDermott. “Learning Midlevel Auditory Codes from Natural Sound Statistics.” Neural Computation, vol. 30, no. 3, Mar. 2018, pp. 631–69. © 2018 Massachusetts Institute of Technologyen_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciencesen_US
dc.contributor.mitauthorMlynarski, Wiktor
dc.contributor.mitauthorMcDermott, Joshua H.
dc.relation.journalNeural Computationen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2018-02-23T19:55:53Z
dspace.orderedauthorsMłynarski, Wiktor; McDermott, Josh H.en_US
dspace.embargo.termsNen_US
dc.identifier.orcidhttps://orcid.org/0000-0002-3791-5656
dc.identifier.orcidhttps://orcid.org/0000-0002-3965-2503
mit.licensePUBLISHER_POLICYen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record