Return to search

Hierarchy and invariance in auditory cortical computation

Thesis: Ph. D. in Neuroscience, Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, June, 2019 / Cataloged from the PDF version of thesis. "June 2019"--Hand written on title page. / Includes bibliographical references. / With ease, we recognize a friend's voice in a crowd, or pick out the first violin in a concerto. But the effortlessness of everyday perception masks its computational challenge. Perception does not occur in the eyes and ears - indeed, nearly half of primate cortex is dedicated to it. While much is known about peripheral auditory processing, auditory cortex remains poorly understood. This thesis addresses basic questions about the functional and computational organization of human auditory cortex through three studies. In the first study we show that a hierarchical neural network model optimized to recognize speech and music does so at human levels, exhibits a similar pattern of behavioral errors, and predicts cortical responses, as measured with fMRI. The multi-task optimization procedure we introduce produces separate music and speech pathways after a shared front end, potentially recapitulating aspects of auditory cortical functional organization. Within the model, different layers best predict primary and non-primary voxels, revealing a hierarchical organization in human auditory cortex. We then seek to characterize the representational transformations that occur across stages of the putative cortical hierarchy, probing for one candidate: invariance to realworld background noise. To measure invariance, we correlate voxel responses to natural sounds with and without real-world background noise. Non-primary responses are substantially more noise-invariant than primary responses. These results illustrate a representational consequence of the potential hierarchical organization of the auditory system. Lastly, we explore of the generality of deep neural networks as models of human hearing by simulating many psychophysical and fMRI experiments on the above-described neural network model. The results provide an extensive comparison of the performance characteristics and internal representations of a deep neural network with those of humans. We observe many similarities that suggest that the model replicates a broad variety of aspects of auditory perception. However, we also find discrepancies that suggest targets for future modeling efforts. / by Alexander James Eaton Kell. / Ph. D. in Neuroscience / Ph.D.inNeuroscience Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences

Identiferoai:union.ndltd.org:MIT/oai:dspace.mit.edu:1721.1/132746
Date January 2019
CreatorsKell, Alexander James Eaton.
ContributorsMassachusetts Institute of Technology. Department of Brain and Cognitive Sciences., Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences
PublisherMassachusetts Institute of Technology
Source SetsM.I.T. Theses and Dissertation
LanguageEnglish
Detected LanguageEnglish
TypeThesis
Format236 pages, application/pdf
RightsMIT theses may be protected by copyright. Please reuse MIT thesis content according to the MIT Libraries Permissions Policy, which is available through the URL provided., http://dspace.mit.edu/handle/1721.1/7582

Page generated in 0.0358 seconds