1
|
Norman-Haignere SV, Keshishian MK, Devinsky O, Doyle W, McKhann GM, Schevon CA, Flinker A, Mesgarani N. Temporal integration in human auditory cortex is predominantly yoked to absolute time, not structure duration. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.09.23.614358. [PMID: 39386565 PMCID: PMC11463558 DOI: 10.1101/2024.09.23.614358] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 10/12/2024]
Abstract
Sound structures such as phonemes and words have highly variable durations. Thus, there is a fundamental difference between integrating across absolute time (e.g., 100 ms) vs. sound structure (e.g., phonemes). Auditory and cognitive models have traditionally cast neural integration in terms of time and structure, respectively, but the extent to which cortical computations reflect time or structure remains unknown. To answer this question, we rescaled the duration of all speech structures using time stretching/compression and measured integration windows in the human auditory cortex using a new experimental/computational method applied to spatiotemporally precise intracranial recordings. We observed significantly longer integration windows for stretched speech, but this lengthening was very small (∼5%) relative to the change in structure durations, even in non-primary regions strongly implicated in speech-specific processing. These findings demonstrate that time-yoked computations dominate throughout the human auditory cortex, placing important constraints on neurocomputational models of structure processing.
Collapse
|
2
|
Özcan F, Alkan A. Neural decoding of inferior colliculus multiunit activity for sound category identification with temporal correlation and transfer learning. NETWORK (BRISTOL, ENGLAND) 2024; 35:101-133. [PMID: 37982591 DOI: 10.1080/0954898x.2023.2282576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 11/07/2023] [Indexed: 11/21/2023]
Abstract
Natural sounds are easily perceived and identified by humans and animals. Despite this, the neural transformations that enable sound perception remain largely unknown. It is thought that the temporal characteristics of sounds may be reflected in auditory assembly responses at the inferior colliculus (IC) and which may play an important role in identification of natural sounds. In our study, natural sounds will be predicted from multi-unit activity (MUA) signals collected in the IC. Data is obtained from an international platform publicly accessible. The temporal correlation values of the MUA signals are converted into images. We used two different segment sizes and with a denoising method, we generated four subsets for the classification. Using pre-trained convolutional neural networks (CNNs), features of the images were extracted and the type of heard sound was classified. For this, we applied transfer learning from Alexnet, Googlenet and Squeezenet CNNs. The classifiers support vector machines (SVM), k-nearest neighbour (KNN), Naive Bayes and Ensemble were used. The accuracy, sensitivity, specificity, precision and F1 score were measured as evaluation parameters. By using all the tests and removing the noise, the accuracy improved significantly. These results will allow neuroscientists to make interesting conclusions.
Collapse
Affiliation(s)
- Fatma Özcan
- Electrical & Electronics Engineering Department, Kahramanmaras Sutcu Imam University, Kahramanmaraş, Turkey
| | - Ahmet Alkan
- Electrical & Electronics Engineering Department, Kahramanmaras Sutcu Imam University, Kahramanmaraş, Turkey
| |
Collapse
|
3
|
Nocon JC, Gritton HJ, James NM, Mount RA, Qu Z, Han X, Sen K. Parvalbumin neurons enhance temporal coding and reduce cortical noise in complex auditory scenes. Commun Biol 2023; 6:751. [PMID: 37468561 PMCID: PMC10356822 DOI: 10.1038/s42003-023-05126-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2022] [Accepted: 07/10/2023] [Indexed: 07/21/2023] Open
Abstract
Cortical representations supporting many cognitive abilities emerge from underlying circuits comprised of several different cell types. However, cell type-specific contributions to rate and timing-based cortical coding are not well-understood. Here, we investigated the role of parvalbumin neurons in cortical complex scene analysis. Many complex scenes contain sensory stimuli which are highly dynamic in time and compete with stimuli at other spatial locations. Parvalbumin neurons play a fundamental role in balancing excitation and inhibition in cortex and sculpting cortical temporal dynamics; yet their specific role in encoding complex scenes via timing-based coding, and the robustness of temporal representations to spatial competition, has not been investigated. Here, we address these questions in auditory cortex of mice using a cocktail party-like paradigm, integrating electrophysiology, optogenetic manipulations, and a family of spike-distance metrics, to dissect parvalbumin neurons' contributions towards rate and timing-based coding. We find that suppressing parvalbumin neurons degrades cortical discrimination of dynamic sounds in a cocktail party-like setting via changes in rapid temporal modulations in rate and spike timing, and over a wide range of time-scales. Our findings suggest that parvalbumin neurons play a critical role in enhancing cortical temporal coding and reducing cortical noise, thereby improving representations of dynamic stimuli in complex scenes.
Collapse
Affiliation(s)
- Jian Carlo Nocon
- Neurophotonics Center, Boston University, Boston, 02215, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, 02215, MA, USA
- Hearing Research Center, Boston University, Boston, 02215, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, 02215, MA, USA
| | - Howard J Gritton
- Department of Comparative Biosciences, University of Illinois, Urbana, 61820, IL, USA
- Department of Bioengineering, University of Illinois, Urbana, 61820, IL, USA
| | - Nicholas M James
- Neurophotonics Center, Boston University, Boston, 02215, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, 02215, MA, USA
- Hearing Research Center, Boston University, Boston, 02215, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, 02215, MA, USA
| | - Rebecca A Mount
- Neurophotonics Center, Boston University, Boston, 02215, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, 02215, MA, USA
- Hearing Research Center, Boston University, Boston, 02215, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, 02215, MA, USA
| | - Zhili Qu
- Department of Comparative Biosciences, University of Illinois, Urbana, 61820, IL, USA
- Department of Bioengineering, University of Illinois, Urbana, 61820, IL, USA
| | - Xue Han
- Neurophotonics Center, Boston University, Boston, 02215, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, 02215, MA, USA
- Hearing Research Center, Boston University, Boston, 02215, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, 02215, MA, USA
| | - Kamal Sen
- Neurophotonics Center, Boston University, Boston, 02215, MA, USA.
- Center for Systems Neuroscience, Boston University, Boston, 02215, MA, USA.
- Hearing Research Center, Boston University, Boston, 02215, MA, USA.
- Department of Biomedical Engineering, Boston University, Boston, 02215, MA, USA.
| |
Collapse
|
4
|
Barzelay O, David S, Delgutte B. Effect of Reverberation on Neural Responses to Natural Speech in Rabbit Auditory Midbrain: No Evidence for a Neural Dereverberation Mechanism. eNeuro 2023; 10:ENEURO.0447-22.2023. [PMID: 37072174 PMCID: PMC10179871 DOI: 10.1523/eneuro.0447-22.2023] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2022] [Revised: 02/10/2023] [Accepted: 02/22/2023] [Indexed: 04/20/2023] Open
Abstract
Reverberation is ubiquitous in everyday acoustic environments. It degrades both binaural cues and the envelope modulations of sounds and thus can impair speech perception. Still, both humans and animals can accurately perceive reverberant stimuli in most everyday settings. Previous neurophysiological and perceptual studies have suggested the existence of neural mechanisms that partially compensate for the effects of reverberation. However, these studies were limited by their use of either highly simplified stimuli or rudimentary reverberation simulations. To further characterize how reverberant stimuli are processed by the auditory system, we recorded single-unit (SU) and multiunit (MU) activity from the inferior colliculus (IC) of unanesthetized rabbits in response to natural speech utterances presented with no reverberation ("dry") and in various degrees of simulated reverberation (direct-to-reverberant energy ratios (DRRs) ranging from 9.4 to -8.2 dB). Linear stimulus reconstruction techniques (Mesgarani et al., 2009) were used to quantify the amount of speech information available in the responses of neural ensembles. We found that high-quality spectrogram reconstructions could be obtained for dry speech and in moderate reverberation from ensembles of 25 units. However, spectrogram reconstruction quality deteriorated in severe reverberation for both MUs and SUs such that the neural degradation paralleled the degradation in the stimulus spectrogram. Furthermore, spectrograms reconstructed from responses to reverberant stimuli resembled spectrograms of reverberant speech better than spectrograms of dry speech. Overall, the results provide no evidence for a dereverberation mechanism in neural responses from the rabbit IC when studied with linear reconstruction techniques.
Collapse
Affiliation(s)
- Oded Barzelay
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, MA 02114-3096
- Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, MA 02115
| | - Stephen David
- Oregon Research Hearing Center, Oregon Health and Science University, Portland, OR 97239-3098
| | - Bertrand Delgutte
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear, Boston, MA 02114-3096
- Department of Otolaryngology, Head and Neck Surgery, Harvard Medical School, Boston, MA 02115
| |
Collapse
|
5
|
He F, Stevenson IH, Escabí MA. Two stages of bandwidth scaling drives efficient neural coding of natural sounds. PLoS Comput Biol 2023; 19:e1010862. [PMID: 36787338 PMCID: PMC9970106 DOI: 10.1371/journal.pcbi.1010862] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2022] [Revised: 02/27/2023] [Accepted: 01/09/2023] [Indexed: 02/15/2023] Open
Abstract
Theories of efficient coding propose that the auditory system is optimized for the statistical structure of natural sounds, yet the transformations underlying optimal acoustic representations are not well understood. Using a database of natural sounds including human speech and a physiologically-inspired auditory model, we explore the consequences of peripheral (cochlear) and mid-level (auditory midbrain) filter tuning transformations on the representation of natural sound spectra and modulation statistics. Whereas Fourier-based sound decompositions have constant time-frequency resolution at all frequencies, cochlear and auditory midbrain filters bandwidths increase proportional to the filter center frequency. This form of bandwidth scaling produces a systematic decrease in spectral resolution and increase in temporal resolution with increasing frequency. Here we demonstrate that cochlear bandwidth scaling produces a frequency-dependent gain that counteracts the tendency of natural sound power to decrease with frequency, resulting in a whitened output representation. Similarly, bandwidth scaling in mid-level auditory filters further enhances the representation of natural sounds by producing a whitened modulation power spectrum (MPS) with higher modulation entropy than both the cochlear outputs and the conventional Fourier MPS. These findings suggest that the tuning characteristics of the peripheral and mid-level auditory system together produce a whitened output representation in three dimensions (frequency, temporal and spectral modulation) that reduces redundancies and allows for a more efficient use of neural resources. This hierarchical multi-stage tuning strategy is thus likely optimized to extract available information and may underlies perceptual sensitivity to natural sounds.
Collapse
Affiliation(s)
- Fengrong He
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Ian H. Stevenson
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- The Connecticut Institute for Brain and Cognitive Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- The Connecticut Institute for Brain and Cognitive Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- * E-mail:
| |
Collapse
|
6
|
Norman-Haignere SV, Long LK, Devinsky O, Doyle W, Irobunda I, Merricks EM, Feldstein NA, McKhann GM, Schevon CA, Flinker A, Mesgarani N. Multiscale temporal integration organizes hierarchical computation in human auditory cortex. Nat Hum Behav 2022; 6:455-469. [PMID: 35145280 PMCID: PMC8957490 DOI: 10.1038/s41562-021-01261-y] [Citation(s) in RCA: 25] [Impact Index Per Article: 12.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2020] [Accepted: 11/18/2021] [Indexed: 01/11/2023]
Abstract
To derive meaning from sound, the brain must integrate information across many timescales. What computations underlie multiscale integration in human auditory cortex? Evidence suggests that auditory cortex analyses sound using both generic acoustic representations (for example, spectrotemporal modulation tuning) and category-specific computations, but the timescales over which these putatively distinct computations integrate remain unclear. To answer this question, we developed a general method to estimate sensory integration windows-the time window when stimuli alter the neural response-and applied our method to intracranial recordings from neurosurgical patients. We show that human auditory cortex integrates hierarchically across diverse timescales spanning from ~50 to 400 ms. Moreover, we find that neural populations with short and long integration windows exhibit distinct functional properties: short-integration electrodes (less than ~200 ms) show prominent spectrotemporal modulation selectivity, while long-integration electrodes (greater than ~200 ms) show prominent category selectivity. These findings reveal how multiscale integration organizes auditory computation in the human brain.
Collapse
Affiliation(s)
- Sam V Norman-Haignere
- Zuckerman Mind, Brain, Behavior Institute, Columbia University,HHMI Postdoctoral Fellow of the Life Sciences Research Foundation
| | - Laura K. Long
- Zuckerman Mind, Brain, Behavior Institute, Columbia University,Doctoral Program in Neurobiology and Behavior, Columbia University
| | - Orrin Devinsky
- Department of Neurology, NYU Langone Medical Center,Comprehensive Epilepsy Center, NYU Langone Medical Center
| | - Werner Doyle
- Comprehensive Epilepsy Center, NYU Langone Medical Center,Department of Neurosurgery, NYU Langone Medical Center
| | - Ifeoma Irobunda
- Department of Neurology, Columbia University Irving Medical Center
| | | | - Neil A. Feldstein
- Department of Neurological Surgery, Columbia University Irving Medical Center
| | - Guy M. McKhann
- Department of Neurological Surgery, Columbia University Irving Medical Center
| | | | - Adeen Flinker
- Department of Neurology, NYU Langone Medical Center,Comprehensive Epilepsy Center, NYU Langone Medical Center,Department of Biomedical Engineering, NYU Tandon School of Engineering
| | - Nima Mesgarani
- Zuckerman Mind, Brain, Behavior Institute, Columbia University,Doctoral Program in Neurobiology and Behavior, Columbia University,Department of Electrical Engineering, Columbia University
| |
Collapse
|
7
|
Gentile Polese A, Nigam S, Hurley LM. 5-HT1A Receptors Alter Temporal Responses to Broadband Vocalizations in the Mouse Inferior Colliculus Through Response Suppression. Front Neural Circuits 2021; 15:718348. [PMID: 34512276 PMCID: PMC8430226 DOI: 10.3389/fncir.2021.718348] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2021] [Accepted: 07/19/2021] [Indexed: 01/21/2023] Open
Abstract
Neuromodulatory systems may provide information on social context to auditory brain regions, but relatively few studies have assessed the effects of neuromodulation on auditory responses to acoustic social signals. To address this issue, we measured the influence of the serotonergic system on the responses of neurons in a mouse auditory midbrain nucleus, the inferior colliculus (IC), to vocal signals. Broadband vocalizations (BBVs) are human-audible signals produced by mice in distress as well as by female mice in opposite-sex interactions. The production of BBVs is context-dependent in that they are produced both at early stages of interactions as females physically reject males and at later stages as males mount females. Serotonin in the IC of males corresponds to these events, and is elevated more in males that experience less female rejection. We measured the responses of single IC neurons to five recorded examples of BBVs in anesthetized mice. We then locally activated the 5-HT1A receptor through iontophoretic application of 8-OH-DPAT. IC neurons showed little selectivity for different BBVs, but spike trains were characterized by local regions of high spike probability, which we called "response features." Response features varied across neurons and also across calls for individual neurons, ranging from 1 to 7 response features for responses of single neurons to single calls. 8-OH-DPAT suppressed spikes and also reduced the numbers of response features. The weakest response features were the most likely to disappear, suggestive of an "iceberg"-like effect in which activation of the 5-HT1A receptor suppressed weakly suprathreshold response features below the spiking threshold. Because serotonin in the IC is more likely to be elevated for mounting-associated BBVs than for rejection-associated BBVs, these effects of the 5-HT1A receptor could contribute to the differential auditory processing of BBVs in different behavioral subcontexts.
Collapse
Affiliation(s)
- Arianna Gentile Polese
- Department of Cell and Developmental Biology, University of Colorado Anschutz Medical Campus, Aurora, CO, United States
- Department of Biology, Program in Neuroscience, Indiana University Bloomington, Bloomington, IN, United States
| | - Sunny Nigam
- Department of Neurobiology and Anatomy, McGovern Medical School, The University of Texas Health Science Center at Houston, Houston, TX, United States
- Department of Physics, Indiana University Bloomington, Bloomington, IN, United States
| | - Laura M. Hurley
- Department of Neurobiology and Anatomy, McGovern Medical School, The University of Texas Health Science Center at Houston, Houston, TX, United States
| |
Collapse
|
8
|
Distinct neural ensemble response statistics are associated with recognition and discrimination of natural sound textures. Proc Natl Acad Sci U S A 2020; 117:31482-31493. [PMID: 33219122 DOI: 10.1073/pnas.2005644117] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
The perception of sound textures, a class of natural sounds defined by statistical sound structure such as fire, wind, and rain, has been proposed to arise through the integration of time-averaged summary statistics. Where and how the auditory system might encode these summary statistics to create internal representations of these stationary sounds, however, is unknown. Here, using natural textures and synthetic variants with reduced statistics, we show that summary statistics modulate the correlations between frequency organized neuron ensembles in the awake rabbit inferior colliculus (IC). These neural ensemble correlation statistics capture high-order sound structure and allow for accurate neural decoding in a single trial recognition task with evidence accumulation times approaching 1 s. In contrast, the average activity across the neural ensemble (neural spectrum) provides a fast (tens of milliseconds) and salient signal that contributes primarily to texture discrimination. Intriguingly, perceptual studies in human listeners reveal analogous trends: the sound spectrum is integrated quickly and serves as a salient discrimination cue while high-order sound statistics are integrated slowly and contribute substantially more toward recognition. The findings suggest statistical sound cues such as the sound spectrum and correlation structure are represented by distinct response statistics in auditory midbrain ensembles, and that these neural response statistics may have dissociable roles and time scales for the recognition and discrimination of natural sounds.
Collapse
|
9
|
Spiking network optimized for word recognition in noise predicts auditory system hierarchy. PLoS Comput Biol 2020; 16:e1007558. [PMID: 32559204 PMCID: PMC7329140 DOI: 10.1371/journal.pcbi.1007558] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2019] [Revised: 07/01/2020] [Accepted: 11/22/2019] [Indexed: 11/21/2022] Open
Abstract
The auditory neural code is resilient to acoustic variability and capable of recognizing sounds amongst competing sound sources, yet, the transformations enabling noise robust abilities are largely unknown. We report that a hierarchical spiking neural network (HSNN) optimized to maximize word recognition accuracy in noise and multiple talkers predicts organizational hierarchy of the ascending auditory pathway. Comparisons with data from auditory nerve, midbrain, thalamus and cortex reveals that the optimal HSNN predicts several transformations of the ascending auditory pathway including a sequential loss of temporal resolution and synchronization ability, increasing sparseness, and selectivity. The optimal organizational scheme enhances performance by selectively filtering out noise and fast temporal cues such as voicing periodicity, that are not directly relevant to the word recognition task. An identical network arranged to enable high information transfer fails to predict auditory pathway organization and has substantially poorer performance. Furthermore, conventional single-layer linear and nonlinear receptive field networks that capture the overall feature extraction of the HSNN fail to achieve similar performance. The findings suggest that the auditory pathway hierarchy and its sequential nonlinear feature extraction computations enhance relevant cues while removing non-informative sources of noise, thus enhancing the representation of sounds in noise impoverished conditions. The brain’s ability to recognize sounds in the presence of competing sounds or background noise is essential for everyday hearing tasks. How the brain accomplishes noise resiliency, however, is poorly understood. Using neural recordings from the ascending auditory pathway and an auditory spiking network model trained for sound recognition in noise we explore the computational strategies that enable noise robustness. Our results suggest that the hierarchical feature organization of the ascending auditory pathway and the resulting computations are critical for sound recognition in the presence of noise.
Collapse
|
10
|
Sadeghi M, Zhai X, Stevenson IH, Escabí MA. A neural ensemble correlation code for sound category identification. PLoS Biol 2019; 17:e3000449. [PMID: 31574079 PMCID: PMC6788721 DOI: 10.1371/journal.pbio.3000449] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2019] [Revised: 10/11/2019] [Accepted: 09/03/2019] [Indexed: 12/25/2022] Open
Abstract
Humans and other animals effortlessly identify natural sounds and categorize them into behaviorally relevant categories. Yet, the acoustic features and neural transformations that enable sound recognition and the formation of perceptual categories are largely unknown. Here, using multichannel neural recordings in the auditory midbrain of unanesthetized female rabbits, we first demonstrate that neural ensemble activity in the auditory midbrain displays highly structured correlations that vary with distinct natural sound stimuli. These stimulus-driven correlations can be used to accurately identify individual sounds using single-response trials, even when the sounds do not differ in their spectral content. Combining neural recordings and an auditory model, we then show how correlations between frequency-organized auditory channels can contribute to discrimination of not just individual sounds but sound categories. For both the model and neural data, spectral and temporal correlations achieved similar categorization performance and appear to contribute equally. Moreover, both the neural and model classifiers achieve their best task performance when they accumulate evidence over a time frame of approximately 1-2 seconds, mirroring human perceptual trends. These results together suggest that time-frequency correlations in sounds may be reflected in the correlations between auditory midbrain ensembles and that these correlations may play an important role in the identification and categorization of natural sounds.
Collapse
Affiliation(s)
- Mina Sadeghi
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Xiu Zhai
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Ian H. Stevenson
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- * E-mail:
| |
Collapse
|
11
|
Chen C, Read HL, Escabí MA. A temporal integration mechanism enhances frequency selectivity of broadband inputs to inferior colliculus. PLoS Biol 2019; 17:e2005861. [PMID: 31233489 PMCID: PMC6611646 DOI: 10.1371/journal.pbio.2005861] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 07/05/2019] [Accepted: 05/22/2019] [Indexed: 11/18/2022] Open
Abstract
Accurately resolving frequency components in sounds is essential for sound recognition, yet there is little direct evidence for how frequency selectivity is preserved or newly created across auditory structures. We demonstrate that prepotentials (PPs) with physiological properties resembling presynaptic potentials from broadly tuned brainstem inputs can be recorded concurrently with postsynaptic action potentials in inferior colliculus (IC). These putative brainstem inputs (PBIs) are broadly tuned and exhibit delayed and spectrally interleaved excitation and inhibition not present in the simultaneously recorded IC neurons (ICNs). A sharpening of tuning is accomplished locally at the expense of spike-timing precision through nonlinear temporal integration of broadband inputs. A neuron model replicates the finding and demonstrates that temporal integration alone can degrade timing precision while enhancing frequency tuning through interference of spectrally in- and out-of-phase inputs. These findings suggest that, in contrast to current models that require local inhibition, frequency selectivity can be sharpened through temporal integration, thus supporting an alternative computational strategy to quickly refine frequency selectivity.
Collapse
Affiliation(s)
- Chen Chen
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Heather L. Read
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| |
Collapse
|
12
|
Luo J, Macias S, Ness TV, Einevoll GT, Zhang K, Moss CF. Neural timing of stimulus events with microsecond precision. PLoS Biol 2018; 16:e2006422. [PMID: 30365484 PMCID: PMC6221347 DOI: 10.1371/journal.pbio.2006422] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2018] [Revised: 11/07/2018] [Accepted: 10/10/2018] [Indexed: 12/29/2022] Open
Abstract
Temporal analysis of sound is fundamental to auditory processing throughout the animal kingdom. Echolocating bats are powerful models for investigating the underlying mechanisms of auditory temporal processing, as they show microsecond precision in discriminating the timing of acoustic events. However, the neural basis for microsecond auditory discrimination in bats has eluded researchers for decades. Combining extracellular recordings in the midbrain inferior colliculus (IC) and mathematical modeling, we show that microsecond precision in registering stimulus events emerges from synchronous neural firing, revealed through low-latency variability of stimulus-evoked extracellular field potentials (EFPs, 200–600 Hz). The temporal precision of the EFP increases with the number of neurons firing in synchrony. Moreover, there is a functional relationship between the temporal precision of the EFP and the spectrotemporal features of the echolocation calls. In addition, EFP can measure the time difference of simulated echolocation call–echo pairs with microsecond precision. We propose that synchronous firing of populations of neurons operates in diverse species to support temporal analysis for auditory localization and complex sound processing. We routinely rely on a stopwatch to precisely measure the time it takes for an athlete to reach the finish line. Without the assistance of such a timing device, our measurement of elapsed time becomes imprecise. By contrast, some animals, such as echolocating bats, naturally perform timing tasks with remarkable precision. Behavioral research has shown that echolocating bats can estimate the elapsed time between sonar cries and echo returns with a precision in the range of microseconds. However, the neural basis for such microsecond precision has remained a puzzle to scientists. Combining extracellular recordings in the bat’s inferior colliculus (IC)—a midbrain nucleus of the auditory pathway—and mathematical modeling, we show that microsecond precision in registering stimulus events emerges from synchronous neural firing. Our recordings revealed a low-latency variability of stimulus-evoked extracellular field potentials (EFPs), which, according to our mathematical modeling, was determined by the number of firing neurons and their synchrony. Moreover, the acoustic features of echolocation calls, such as signal duration and bandwidth, which the bat dynamically modulates during prey capture, also modulate the precision of EFPs. These findings have broad implications for understanding temporal analysis of acoustic signals in a wide range of auditory behaviors across the animal kingdom.
Collapse
Affiliation(s)
- Jinhong Luo
- Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, Maryland, United States of America
- * E-mail: (JL); (CFM)
| | - Silvio Macias
- Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, Maryland, United States of America
| | - Torbjørn V. Ness
- Faculty of Science and Technology, Norwegian University of Life Sciences, Ås, Norway
| | - Gaute T. Einevoll
- Faculty of Science and Technology, Norwegian University of Life Sciences, Ås, Norway
- Department of Physics, University of Oslo, Oslo, Norway
| | - Kechen Zhang
- Department of Biomedical Engineering, Johns Hopkins School of Medicine, Baltimore, Maryland, United States of America
| | - Cynthia F. Moss
- Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, Maryland, United States of America
- * E-mail: (JL); (CFM)
| |
Collapse
|
13
|
A Hierarchy of Time Scales for Discriminating and Classifying the Temporal Shape of Sound in Three Auditory Cortical Fields. J Neurosci 2018; 38:6967-6982. [PMID: 29954851 DOI: 10.1523/jneurosci.2871-17.2018] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2017] [Revised: 05/29/2018] [Accepted: 06/17/2018] [Indexed: 11/21/2022] Open
Abstract
Auditory cortex is essential for mammals, including rodents, to detect temporal "shape" cues in the sound envelope but it remains unclear how different cortical fields may contribute to this ability (Lomber and Malhotra, 2008; Threlkeld et al., 2008). Previously, we found that precise spiking patterns provide a potential neural code for temporal shape cues in the sound envelope in the primary auditory (A1), and ventral auditory field (VAF) and caudal suprarhinal auditory field (cSRAF) of the rat (Lee et al., 2016). Here, we extend these findings and characterize the time course of the temporally precise output of auditory cortical neurons in male rats. A pairwise sound discrimination index and a Naive Bayesian classifier are used to determine how these spiking patterns could provide brain signals for behavioral discrimination and classification of sounds. We find response durations and optimal time constants for discriminating sound envelope shape increase in rank order with: A1 < VAF < cSRAF. Accordingly, sustained spiking is more prominent and results in more robust sound discrimination in non-primary cortex versus A1. Spike-timing patterns classify 10 different sound envelope shape sequences and there is a twofold increase in maximal performance when pooling output across the neuron population indicating a robust distributed neural code in all three cortical fields. Together, these results support the idea that temporally precise spiking patterns from primary and non-primary auditory cortical fields provide the necessary signals for animals to discriminate and classify a large range of temporal shapes in the sound envelope.SIGNIFICANCE STATEMENT Functional hierarchies in the visual cortices support the concept that classification of visual objects requires successive cortical stages of processing including a progressive increase in classical receptive field size. The present study is significant as it supports the idea that a similar progression exists in auditory cortices in the time domain. We demonstrate for the first time that three cortices provide temporal spiking patterns for robust temporal envelope shape discrimination but only the ventral non-primary cortices do so on long time scales. This study raises the possibility that primary and non-primary cortices provide unique temporal spiking patterns and time scales for perception of sound envelope shape.
Collapse
|
14
|
Lee CM, Osman AF, Volgushev M, Escabí MA, Read HL. Neural spike-timing patterns vary with sound shape and periodicity in three auditory cortical fields. J Neurophysiol 2016; 115:1886-904. [PMID: 26843599 DOI: 10.1152/jn.00784.2015] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2015] [Accepted: 01/29/2016] [Indexed: 11/22/2022] Open
Abstract
Mammals perceive a wide range of temporal cues in natural sounds, and the auditory cortex is essential for their detection and discrimination. The rat primary (A1), ventral (VAF), and caudal suprarhinal (cSRAF) auditory cortical fields have separate thalamocortical pathways that may support unique temporal cue sensitivities. To explore this, we record responses of single neurons in the three fields to variations in envelope shape and modulation frequency of periodic noise sequences. Spike rate, relative synchrony, and first-spike latency metrics have previously been used to quantify neural sensitivities to temporal sound cues; however, such metrics do not measure absolute spike timing of sustained responses to sound shape. To address this, in this study we quantify two forms of spike-timing precision, jitter, and reliability. In all three fields, we find that jitter decreases logarithmically with increase in the basis spline (B-spline) cutoff frequency used to shape the sound envelope. In contrast, reliability decreases logarithmically with increase in sound envelope modulation frequency. In A1, jitter and reliability vary independently, whereas in ventral cortical fields, jitter and reliability covary. Jitter time scales increase (A1 < VAF < cSRAF) and modulation frequency upper cutoffs decrease (A1 > VAF > cSRAF) with ventral progression from A1. These results suggest a transition from independent encoding of shape and periodicity sound cues on short time scales in A1 to a joint encoding of these same cues on longer time scales in ventral nonprimary cortices.
Collapse
Affiliation(s)
- Christopher M Lee
- Department of Psychology, University of Connecticut, Storrs, Connecticut
| | - Ahmad F Osman
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and
| | - Maxim Volgushev
- Department of Psychology, University of Connecticut, Storrs, Connecticut
| | - Monty A Escabí
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut
| | - Heather L Read
- Department of Psychology, University of Connecticut, Storrs, Connecticut; Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and
| |
Collapse
|
15
|
Lyzwa D, Herrmann JM, Wörgötter F. Natural Vocalizations in the Mammalian Inferior Colliculus are Broadly Encoded by a Small Number of Independent Multi-Units. Front Neural Circuits 2016; 9:91. [PMID: 26869890 PMCID: PMC4740783 DOI: 10.3389/fncir.2015.00091] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2015] [Accepted: 12/28/2015] [Indexed: 11/18/2022] Open
Abstract
How complex natural sounds are represented by the main converging center of the auditory midbrain, the central inferior colliculus, is an open question. We applied neural discrimination to determine the variation of detailed encoding of individual vocalizations across the best frequency gradient of the central inferior colliculus. The analysis was based on collective responses from several neurons. These multi-unit spike trains were recorded from guinea pigs exposed to a spectrotemporally rich set of eleven species-specific vocalizations. Spike trains of disparate units from the same recording were combined in order to investigate whether groups of multi-unit clusters represent the whole set of vocalizations more reliably than only one unit, and whether temporal response correlations between them facilitate an unambiguous neural representation of the vocalizations. We found a spatial distribution of the capability to accurately encode groups of vocalizations across the best frequency gradient. Different vocalizations are optimally discriminated at different locations of the best frequency gradient. Furthermore, groups of a few multi-unit clusters yield improved discrimination over only one multi-unit cluster between all tested vocalizations. However, temporal response correlations between units do not yield better discrimination. Our study is based on a large set of units of simultaneously recorded responses from several guinea pigs and electrode insertion positions. Our findings suggest a broadly distributed code for behaviorally relevant vocalizations in the mammalian inferior colliculus. Responses from a few non-interacting units are sufficient to faithfully represent the whole set of studied vocalizations with diverse spectrotemporal properties.
Collapse
Affiliation(s)
- Dominika Lyzwa
- Max Planck Institute for Dynamics and Self-OrganizationGöttingen, Germany
- Institute for Nonlinear Dynamics, Physics Department, Georg-August-UniversityGöttingen, Germany
- Bernstein Focus NeurotechnologyGöttingen, Germany
| | - J. Michael Herrmann
- Bernstein Focus NeurotechnologyGöttingen, Germany
- Institute of Perception, Action and Behavior, School of Informatics, University of EdinburghEdinburgh, UK
| | - Florentin Wörgötter
- Bernstein Focus NeurotechnologyGöttingen, Germany
- Institute for Physics - Biophysics, Georg-August-UniversityGöttingen, Germany
| |
Collapse
|
16
|
Ono M, Ito T. Functional organization of the mammalian auditory midbrain. J Physiol Sci 2015; 65:499-506. [PMID: 26362672 PMCID: PMC10718034 DOI: 10.1007/s12576-015-0394-3] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2015] [Accepted: 08/22/2015] [Indexed: 12/12/2022]
Abstract
The inferior colliculus (IC) is a critical nexus between the auditory brainstem and the forebrain. Parallel auditory pathways that emerge from the brainstem are integrated in the IC. In this integration, de-novo auditory information processed as local and ascending inputs converge via the complex neural circuit of the IC. However, it is still unclear how information is processed within the neural circuit. The purpose of this review is to give an anatomical and physiological overview of the IC neural circuit. We address the functional organization of the IC where the excitatory and inhibitory synaptic inputs interact to shape the responses of IC neurons to sound.
Collapse
Affiliation(s)
- Munenori Ono
- Department of Neuroscience, University of Connecticut Health Center, Farmington, CT, 06030-3401, USA.
- Department of Physiology, School of Medicine, Kanazawa Medical University, Uchinada, Ishikawa, 920-0293, Japan.
| | - Tetsufumi Ito
- Department of Anatomy, Faculty of Medical Sciences, University of Fukui, Eiheiji, Fukui, 910-1193, Japan
- Research and Education Program for Life Science, University of Fukui, Fukui, Fukui, 910-8507, Japan
| |
Collapse
|
17
|
King J, Insanally M, Jin M, Martins ARO, D'amour JA, Froemke RC. Rodent auditory perception: Critical band limitations and plasticity. Neuroscience 2015; 296:55-65. [PMID: 25827498 DOI: 10.1016/j.neuroscience.2015.03.053] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2014] [Revised: 03/20/2015] [Accepted: 03/22/2015] [Indexed: 10/23/2022]
Abstract
What do animals hear? While it remains challenging to adequately assess sensory perception in animal models, it is important to determine perceptual abilities in model systems to understand how physiological processes and plasticity relate to perception, learning, and cognition. Here we discuss hearing in rodents, reviewing previous and recent behavioral experiments querying acoustic perception in rats and mice, and examining the relation between behavioral data and electrophysiological recordings from the central auditory system. We focus on measurements of critical bands, which are psychoacoustic phenomena that seem to have a neural basis in the functional organization of the cochlea and the inferior colliculus. We then discuss how behavioral training, brain stimulation, and neuropathology impact auditory processing and perception.
Collapse
Affiliation(s)
- J King
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA
| | - M Insanally
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA
| | - M Jin
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA
| | - A R O Martins
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA; PhD Programme in Experimental Biology and Biomedicine, Center for Neurosciences and Cell Biology, University of Coimbra, Portugal
| | - J A D'amour
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA
| | - R C Froemke
- Skirball Institute for Biomolecular Medicine, New York University School of Medicine, New York, NY, USA; Department of Otolaryngology, New York University School of Medicine, New York, NY, USA; Department of Neuroscience and Physiology, New York University School of Medicine, New York, NY, USA; Center for Neural Science, New York University, New York, NY, USA.
| |
Collapse
|
18
|
Abstract
To understand the strategies used by the brain to analyze complex environments, we must first characterize how the features of sensory stimuli are encoded in the spiking of neuronal populations. Characterizing a population code requires identifying the temporal precision of spiking and the extent to which spiking is correlated, both between cells and over time. In this study, we characterize the population code for speech in the gerbil inferior colliculus (IC), the hub of the auditory system where inputs from parallel brainstem pathways are integrated for transmission to the cortex. We find that IC spike trains can carry information about speech with sub-millisecond precision, and, consequently, that the temporal correlations imposed by refractoriness can play a significant role in shaping spike patterns. We also find that, in contrast to most other brain areas, the noise correlations between IC cells are extremely weak, indicating that spiking in the population is conditionally independent. These results demonstrate that the problem of understanding the population coding of speech can be reduced to the problem of understanding the stimulus-driven spiking of individual cells, suggesting that a comprehensive model of the subcortical processing of speech may be attainable in the near future.
Collapse
|
19
|
Abstract
In primary auditory cortex (AI), broadly correlated firing has been commonly observed. In contrast, sharply synchronous firing has rarely been seen and has not been well characterized. Therefore, we examined cat AI local subnetworks using cross-correlation and spectrotemporal receptive field (STRF) analysis for neighboring neurons. Sharply synchronous firing responses were observed predominantly for neurons separated by <150 μm. This high synchrony was independent of layers and was present between all distinguishable cell types. The sharpest synchrony was seen in supragranular layers and between regular spiking units. Synchronous spikes conveyed more stimulus information than nonsynchronous spikes. Neighboring neurons in all layers had similar best frequencies and similar STRFs, with the highest similarity in supragranular and granular layers. Spectral tuning selectivity and latency were only moderately conserved in these local, high-synchrony AI subnetworks. Overall, sharp synchrony is a specific characteristic of fine-scale networks within the AI and local functional processing is well ordered and similar, but not identical, for neighboring neurons of all cell types.
Collapse
|
20
|
Abstract
The encoding of sensory information by populations of cortical neurons forms the basis for perception but remains poorly understood. To understand the constraints of cortical population coding we analyzed neural responses to natural sounds recorded in auditory cortex of primates (Macaca mulatta). We estimated stimulus information while varying the composition and size of the considered population. Consistent with previous reports we found that when choosing subpopulations randomly from the recorded ensemble, the average population information increases steadily with population size. This scaling was explained by a model assuming that each neuron carried equal amounts of information, and that any overlap between the information carried by each neuron arises purely from random sampling within the stimulus space. However, when studying subpopulations selected to optimize information for each given population size, the scaling of information was strikingly different: a small fraction of temporally precise cells carried the vast majority of information. This scaling could be explained by an extended model, assuming that the amount of information carried by individual neurons was highly nonuniform, with few neurons carrying large amounts of information. Importantly, these optimal populations can be determined by a single biophysical marker-the neuron's encoding time scale-allowing their detection and readout within biologically realistic circuits. These results show that extrapolations of population information based on random ensembles may overestimate the population size required for stimulus encoding, and that sensory cortical circuits may process information using small but highly informative ensembles.
Collapse
|
21
|
Natural image sequences constrain dynamic receptive fields and imply a sparse code. Brain Res 2013; 1536:53-67. [PMID: 23933349 DOI: 10.1016/j.brainres.2013.07.056] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2013] [Revised: 07/28/2013] [Accepted: 07/31/2013] [Indexed: 11/22/2022]
Abstract
In their natural environment, animals experience a complex and dynamic visual scenery. Under such natural stimulus conditions, neurons in the visual cortex employ a spatially and temporally sparse code. For the input scenario of natural still images, previous work demonstrated that unsupervised feature learning combined with the constraint of sparse coding can predict physiologically measured receptive fields of simple cells in the primary visual cortex. This convincingly indicated that the mammalian visual system is adapted to the natural spatial input statistics. Here, we extend this approach to the time domain in order to predict dynamic receptive fields that can account for both spatial and temporal sparse activation in biological neurons. We rely on temporal restricted Boltzmann machines and suggest a novel temporal autoencoding training procedure. When tested on a dynamic multi-variate benchmark dataset this method outperformed existing models of this class. Learning features on a large dataset of natural movies allowed us to model spatio-temporal receptive fields for single neurons. They resemble temporally smooth transformations of previously obtained static receptive fields and are thus consistent with existing theories. A neuronal spike response model demonstrates how the dynamic receptive field facilitates temporal and population sparseness. We discuss the potential mechanisms and benefits of a spatially and temporally sparse representation of natural visual input.
Collapse
|
22
|
Rode T, Hartmann T, Hubka P, Scheper V, Lenarz M, Lenarz T, Kral A, Lim HH. Neural representation in the auditory midbrain of the envelope of vocalizations based on a peripheral ear model. Front Neural Circuits 2013; 7:166. [PMID: 24155694 PMCID: PMC3800787 DOI: 10.3389/fncir.2013.00166] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2013] [Accepted: 09/24/2013] [Indexed: 11/24/2022] Open
Abstract
The auditory midbrain implant (AMI) consists of a single shank array (20 sites) for stimulation along the tonotopic axis of the central nucleus of the inferior colliculus (ICC) and has been safely implanted in deaf patients who cannot benefit from a cochlear implant (CI). The AMI improves lip-reading abilities and environmental awareness in the implanted patients. However, the AMI cannot achieve the high levels of speech perception possible with the CI. It appears the AMI can transmit sufficient spectral cues but with limited temporal cues required for speech understanding. Currently, the AMI uses a CI-based strategy, which was originally designed to stimulate each frequency region along the cochlea with amplitude-modulated pulse trains matching the envelope of the bandpass-filtered sound components. However, it is unclear if this type of stimulation with only a single site within each frequency lamina of the ICC can elicit sufficient temporal cues for speech perception. At least speech understanding in quiet is still possible with envelope cues as low as 50 Hz. Therefore, we investigated how ICC neurons follow the bandpass-filtered envelope structure of natural stimuli in ketamine-anesthetized guinea pigs. We identified a subset of ICC neurons that could closely follow the envelope structure (up to ß100 Hz) of a diverse set of species-specific calls, which was revealed by using a peripheral ear model to estimate the true bandpass-filtered envelopes observed by the brain. Although previous studies have suggested a complex neural transformation from the auditory nerve to the ICC, our data suggest that the brain maintains a robust temporal code in a subset of ICC neurons matching the envelope structure of natural stimuli. Clinically, these findings suggest that a CI-based strategy may still be effective for the AMI if the appropriate neurons are entrained to the envelope of the acoustic stimulus and can transmit sufficient temporal cues to higher centers.
Collapse
Affiliation(s)
- Thilo Rode
- Department of Otorhinolaryngology, Hannover Medical University Hannover, Germany
| | | | | | | | | | | | | | | |
Collapse
|
23
|
Cortical inhibition reduces information redundancy at presentation of communication sounds in the primary auditory cortex. J Neurosci 2013; 33:10713-28. [PMID: 23804094 DOI: 10.1523/jneurosci.0079-13.2013] [Citation(s) in RCA: 36] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
In all sensory modalities, intracortical inhibition shapes the functional properties of cortical neurons but also influences the responses to natural stimuli. Studies performed in various species have revealed that auditory cortex neurons respond to conspecific vocalizations by temporal spike patterns displaying a high trial-to-trial reliability, which might result from precise timing between excitation and inhibition. Studying the guinea pig auditory cortex, we show that partial blockage of GABAA receptors by gabazine (GBZ) application (10 μm, a concentration that promotes expansion of cortical receptive fields) increased the evoked firing rate and the spike-timing reliability during presentation of communication sounds (conspecific and heterospecific vocalizations), whereas GABAB receptor antagonists [10 μm saclofen; 10-50 μm CGP55845 (p-3-aminopropyl-p-diethoxymethyl phosphoric acid)] had nonsignificant effects. Computing mutual information (MI) from the responses to vocalizations using either the evoked firing rate or the temporal spike patterns revealed that GBZ application increased the MI derived from the activity of single cortical site but did not change the MI derived from population activity. In addition, quantification of information redundancy showed that GBZ significantly increased redundancy at the population level. This result suggests that a potential role of intracortical inhibition is to reduce information redundancy during the processing of natural stimuli.
Collapse
|
24
|
Ranasinghe KG, Vrana WA, Matney CJ, Kilgard MP. Increasing diversity of neural responses to speech sounds across the central auditory pathway. Neuroscience 2013; 252:80-97. [PMID: 23954862 DOI: 10.1016/j.neuroscience.2013.08.005] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2013] [Revised: 07/24/2013] [Accepted: 08/03/2013] [Indexed: 10/26/2022]
Abstract
Neurons at higher stations of each sensory system are responsive to feature combinations not present at lower levels. As a result, the activity of these neurons becomes less redundant than lower levels. We recorded responses to speech sounds from the inferior colliculus and the primary auditory cortex neurons of rats, and tested the hypothesis that primary auditory cortex neurons are more sensitive to combinations of multiple acoustic parameters compared to inferior colliculus neurons. We independently eliminated periodicity information, spectral information and temporal information in each consonant and vowel sound using a noise vocoder. This technique made it possible to test several key hypotheses about speech sound processing. Our results demonstrate that inferior colliculus responses are spatially arranged and primarily determined by the spectral energy and the fundamental frequency of speech, whereas primary auditory cortex neurons generate widely distributed responses to multiple acoustic parameters, and are not strongly influenced by the fundamental frequency of speech. We found no evidence that inferior colliculus or primary auditory cortex was specialized for speech features such as voice onset time or formants. The greater diversity of responses in primary auditory cortex compared to inferior colliculus may help explain how the auditory system can identify a wide range of speech sounds across a wide range of conditions without relying on any single acoustic cue.
Collapse
Affiliation(s)
- K G Ranasinghe
- The University of Texas at Dallas, School of Behavioral Brain Sciences, 800 West Campbell Road, GR41, Richardson, TX 75080-3021, United States.
| | | | | | | |
Collapse
|
25
|
Straka MM, Schendel D, Lim HH. Neural integration and enhancement from the inferior colliculus up to different layers of auditory cortex. J Neurophysiol 2013; 110:1009-20. [PMID: 23719210 DOI: 10.1152/jn.00022.2013] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
While the cochlear implant has successfully restored hearing to many deaf patients, it cannot benefit those without a functional auditory nerve or an implantable cochlea. As an alternative, the auditory midbrain implant (AMI) has been developed and implanted into deaf patients. Consisting of a single-shank array, the AMI is designed for stimulation along the tonotopic gradient of the inferior colliculus (ICC). Although the AMI can provide frequency cues, it appears to insufficiently transmit temporal cues for speech understanding because repeated stimulation of a single site causes strong suppressive and refractory effects. Applying the electrical stimulation to at least two sites within an isofrequency lamina can circumvent these refractory processes. Moreover, coactivation with short intersite delays (<5 ms) can elicit cortical activation which is enhanced beyond the summation of activity induced by the individual sites. The goal of our study was to further investigate the role of the auditory cortex in this enhancement effect. In guinea pigs, we electrically stimulated two locations within an ICC lamina or along different laminae with varying interpulse intervals (0-10 ms) and recorded activity in different locations and layers of primary auditory cortex (A1). Our findings reveal a neural mechanism that integrates activity only from neurons located within the same ICC lamina for short spiking intervals (<6 ms). This mechanism leads to enhanced activity into layers III-V of A1 that is further magnified in supragranular layers. This integration mechanism may contribute to perceptual coding of different sound features that are relevant for improving AMI performance.
Collapse
Affiliation(s)
- Malgorzata M Straka
- Department of Biomedical Engineering, University of Minnesota, Twin Cities, Minneapolis, Minnesota, USA.
| | | | | |
Collapse
|
26
|
Centanni TM, Engineer CT, Kilgard MP. Cortical speech-evoked response patterns in multiple auditory fields are correlated with behavioral discrimination ability. J Neurophysiol 2013; 110:177-89. [PMID: 23596332 DOI: 10.1152/jn.00092.2013] [Citation(s) in RCA: 37] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Different speech sounds evoke unique patterns of activity in primary auditory cortex (A1). Behavioral discrimination by rats is well correlated with the distinctness of the A1 patterns evoked by individual consonants, but only when precise spike timing is preserved. In this study we recorded the speech-evoked responses in the primary, anterior, ventral, and posterior auditory fields of the rat and evaluated whether activity in these fields is better correlated with speech discrimination ability when spike timing information is included or eliminated. Spike timing information improved consonant discrimination in all four of the auditory fields examined. Behavioral discrimination was significantly correlated with neural discrimination in all four auditory fields. The diversity of speech responses across recordings sites was greater in posterior and ventral auditory fields compared with A1 and anterior auditor fields. These results suggest that, while the various auditory fields of the rat process speech sounds differently, neural activity in each field could be used to distinguish between consonant sounds with accuracy that closely parallels behavioral discrimination. Earlier observations in the visual and somatosensory systems that cortical neurons do not rely on spike timing should be reevaluated with more complex natural stimuli to determine whether spike timing contributes to sensory encoding.
Collapse
Affiliation(s)
- T M Centanni
- School of Behavioral and Brain Sciences, University of Texas at Dallas, Richardson, Texas 75080, USA.
| | | | | |
Collapse
|
27
|
Analysis of slow (theta) oscillations as a potential temporal reference frame for information coding in sensory cortices. PLoS Comput Biol 2012; 8:e1002717. [PMID: 23071429 PMCID: PMC3469413 DOI: 10.1371/journal.pcbi.1002717] [Citation(s) in RCA: 69] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2012] [Accepted: 08/12/2012] [Indexed: 11/19/2022] Open
Abstract
While sensory neurons carry behaviorally relevant information in responses that often extend over hundreds of milliseconds, the key units of neural information likely consist of much shorter and temporally precise spike patterns. The mechanisms and temporal reference frames by which sensory networks partition responses into these shorter units of information remain unknown. One hypothesis holds that slow oscillations provide a network-intrinsic reference to temporally partitioned spike trains without exploiting the millisecond-precise alignment of spikes to sensory stimuli. We tested this hypothesis on neural responses recorded in visual and auditory cortices of macaque monkeys in response to natural stimuli. Comparing different schemes for response partitioning revealed that theta band oscillations provide a temporal reference that permits extracting significantly more information than can be obtained from spike counts, and sometimes almost as much information as obtained by partitioning spike trains using precisely stimulus-locked time bins. We further tested the robustness of these partitioning schemes to temporal uncertainty in the decoding process and to noise in the sensory input. This revealed that partitioning using an oscillatory reference provides greater robustness than partitioning using precisely stimulus-locked time bins. Overall, these results provide a computational proof of concept for the hypothesis that slow rhythmic network activity may serve as internal reference frame for information coding in sensory cortices and they foster the notion that slow oscillations serve as key elements for the computations underlying perception.
Collapse
|
28
|
Chen C, Rodriguez FC, Read HL, Escabí MA. Spectrotemporal sound preferences of neighboring inferior colliculus neurons: implications for local circuitry and processing. Front Neural Circuits 2012; 6:62. [PMID: 23060750 PMCID: PMC3461703 DOI: 10.3389/fncir.2012.00062] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2012] [Accepted: 08/19/2012] [Indexed: 11/13/2022] Open
Abstract
How do local circuits in the inferior colliculus (IC) process and transform spectral and temporal sound information? Using a four-tetrode array we examined the functional properties of the IC and metrics of its micro circuitry by recording neural activity from neighboring single neurons in the cat. Spectral and temporal response preferences were compared for neurons found on the same and adjacent tetrodes (ATs), as well as across distant recording sites. We found that neighboring neurons had similar preferences while neurons recorded across distant sites were less similar. Best frequency (BF) was the most correlated parameter between neighboring neurons and BF differences exhibited unique clustering at ~0.3 octave intervals, indicative of the frequency band lamina. Other spectral and temporal parameters of the receptive fields were more similar for neighboring neurons than for those at distant sites and the receptive field similarity was larger for neurons with small differences in BF. Furthermore, correlated firing was stronger for neighboring neuron pairs and increased with proximity and decreasing BF difference. Thus, although response selectivities are quite diverse in the IC, spectral, and temporal preference within a local microcircuit are functionally quite similar. This suggests a scheme where local circuits are organized into zones that are specialized for processing distinct spectrotemporal cues.
Collapse
Affiliation(s)
- Chen Chen
- Department of Electrical and Computer Engineering, University of Connecticut Storrs, CT, USA
| | | | | | | |
Collapse
|