Show simple item record

dc.contributor.authorTuckute, Greta
dc.contributor.authorFeather, Jenelle
dc.contributor.authorBoebinger, Dana
dc.contributor.authorMcDermott, Josh H
dc.date.accessioned2026-04-22T17:50:27Z
dc.date.available2026-04-22T17:50:27Z
dc.date.issued2023-12-13
dc.identifier.urihttps://hdl.handle.net/1721.1/165640
dc.description.abstractModels that predict brain responses to stimuli provide one measure of understanding of a sensory system and have many potential applications in science and engineering. Deep artificial neural networks have emerged as the leading such predictive models of the visual system but are less explored in audition. Prior work provided examples of audio-trained neural networks that produced good predictions of auditory cortical fMRI responses and exhibited correspondence between model stages and brain regions, but left it unclear whether these results generalize to other neural network models and, thus, how to further improve models in this domain. We evaluated model-brain correspondence for publicly available audio neural network models along with in-house models trained on 4 different tasks. Most tested models outpredicted standard spectromporal filter-bank models of auditory cortex and exhibited systematic model-brain correspondence: Middle stages best predicted primary auditory cortex, while deep stages best predicted non-primary cortex. However, some state-of-the-art models produced substantially worse brain predictions. Models trained to recognize speech in background noise produced better brain predictions than models trained to recognize speech in quiet, potentially because hearing in noise imposes constraints on biological auditory representations. The training task influenced the prediction quality for specific cortical tuning properties, with best overall predictions resulting from models trained on multiple tasks. The results generally support the promise of deep neural networks as models of audition, though they also indicate that current models do not explain auditory cortical responses in their entirety.en_US
dc.language.isoen
dc.publisherPublic Library of Science (PLoS)en_US
dc.relation.isversionofhttps://doi.org/10.1371/journal.pbio.3002366en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourcePublic Library of Science (PLoS)en_US
dc.titleMany but not all deep neural network audio models capture brain responses and exhibit correspondence between model stages and brain regionsen_US
dc.typeArticleen_US
dc.identifier.citationTuckute G, Feather J, Boebinger D, McDermott JH (2023) Many but not all deep neural network audio models capture brain responses and exhibit correspondence between model stages and brain regions. PLoS Biol 21(12): e3002366.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Brain and Cognitive Sciencesen_US
dc.contributor.departmentMcGovern Institute for Brain Research at MITen_US
dc.contributor.departmentCenter for Brains, Minds, and Machinesen_US
dc.relation.journalPLOS Biologyen_US
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2026-04-22T17:43:30Z
dspace.orderedauthorsTuckute, G; Feather, J; Boebinger, D; McDermott, JHen_US
dspace.date.submission2026-04-22T17:43:32Z
mit.journal.volume21en_US
mit.journal.issue12en_US
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record