1
|
Mitchell PW, Carney LH. A computational model of auditory chirp-velocity sensitivity and amplitude-modulation tuning in inferior colliculus neurons. J Comput Neurosci 2024; 52:285-302. [PMID: 39259462 DOI: 10.1007/s10827-024-00880-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2024] [Revised: 08/14/2024] [Accepted: 08/17/2024] [Indexed: 09/13/2024]
Abstract
We demonstrate a model of chirp-velocity sensitivity in the inferior colliculus (IC) that retains the tuning to amplitude modulation (AM) that was established in earlier models. The mechanism of velocity sensitivity is sequence detection by octopus cells of the posteroventral cochlear nucleus, which have been proposed in physiological studies to respond preferentially to the order of arrival of cross-frequency inputs of different amplitudes. Model architecture is based on coincidence detection of a combination of excitatory and inhibitory inputs. Chirp-sensitivity of the IC output is largely controlled by the strength and timing of the chirp-sensitive octopus-cell inhibitory input. AM tuning is controlled by inhibition and excitation that are tuned to the same frequency. We present several example neurons that demonstrate the feasibility of the model in simulating realistic chirp-sensitivity and AM tuning for a wide range of characteristic frequencies. Additionally, we explore the systematic impact of varying parameters on model responses. The proposed model can be used to assess the contribution of IC chirp-velocity sensitivity to responses to complex sounds, such as speech.
Collapse
Affiliation(s)
- Paul W Mitchell
- Department of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, NY, USA
| | - Laurel H Carney
- Department of Biomedical Engineering and Neuroscience, University of Rochester, Rochester, NY, USA.
| |
Collapse
|
2
|
Vaziri PA, McDougle SD, Clark DA. Humans use local spectrotemporal correlations to detect rising and falling pitch. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.08.03.606481. [PMID: 39131316 PMCID: PMC11312537 DOI: 10.1101/2024.08.03.606481] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 08/13/2024]
Abstract
To discern speech or appreciate music, the human auditory system detects how pitch increases or decreases over time. However, the algorithms used to detect changes in pitch, or pitch motion, are incompletely understood. Here, using psychophysics, computational modeling, functional neuroimaging, and analysis of recorded speech, we ask if humans detect pitch motion using computations analogous to those used by the visual system. We adapted stimuli from studies of vision to create novel auditory correlated noise stimuli that elicited robust pitch motion percepts. Crucially, these stimuli possess no persistent features across frequency or time, but do possess positive or negative local spectrotemporal correlations in intensity. In psychophysical experiments, we found clear evidence that humans judge pitch direction based on both positive and negative spectrotemporal correlations. The observed sensitivity to negative correlations is a direct analogue of illusory "reverse-phi" motion in vision, and thus constitutes a new auditory illusion. Our behavioral results and computational modeling led us to hypothesize that human auditory processing employs pitch direction opponency. fMRI measurements in auditory cortex supported this hypothesis. To link our psychophysical findings to real-world pitch perception, we analyzed recordings of English and Mandarin speech and discovered that pitch direction was robustly signaled by the same positive and negative spectrotemporal correlations used in our psychophysical tests, suggesting that sensitivity to both positive and negative correlations confers ecological benefits. Overall, this work reveals that motion detection algorithms sensitive to local correlations are deployed by the central nervous system across disparate modalities (vision and audition) and dimensions (space and frequency).
Collapse
Affiliation(s)
| | - Samuel D McDougle
- Dept of Psychology, Yale University, New Haven, CT 06511
- Wu Tsai Institute, Yale University, New Haven, CT 06511
| | - Damon A Clark
- Wu Tsai Institute, Yale University, New Haven, CT 06511
- Dept of Molecular Cellular and Developmental Biology, Yale University, New Haven, CT 06511
- Dept of Physics, Yale University, New Haven, CT 06511
- Dept of Neuroscience, Yale University, New Haven, CT 06511
- Quantitative Biology Institute, Yale University, New Haven, CT 06511
| |
Collapse
|
3
|
Mitchell PW, Carney LH. A Computational Model of Auditory Chirp-Velocity Sensitivity and Amplitude-Modulation Tuning in Inferior Colliculus Neurons. RESEARCH SQUARE 2024:rs.3.rs-4450943. [PMID: 38883707 PMCID: PMC11177976 DOI: 10.21203/rs.3.rs-4450943/v1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2024]
Abstract
We demonstrate a model of chirp-velocity sensitivity in the inferior colliculus (IC) that retains the tuning to amplitude modulation (AM) that was established in earlier models. The mechanism of velocity sensitivity is sequence detection by octopus cells of the posteroventral cochlear nucleus, which have been proposed in physiological studies to respond preferentially to the order of arrival of cross-frequency inputs of different amplitudes. Model architecture is based on coincidence detection of a combination of excitatory and inhibitory inputs. Chirp-sensitivity of the IC output is largely controlled by the strength and timing of the chirp-sensitive octopus-cell inhibitory input. AM tuning is controlled by inhibition and excitation that are tuned to the same frequency. We present several example neurons that demonstrate the feasibility of the model in simulating realistic chirp-sensitivity and AM tuning for a wide range of characteristic frequencies. Additionally, we explore the systematic impact of varying parameters on model responses. The proposed model can be used to assess the contribution of IC chirp-velocity sensitivity to responses to complex sounds, such as speech.
Collapse
Affiliation(s)
- Paul W. Mitchell
- Department of Biomedical Engineering, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
| | - Laurel H. Carney
- Department of Biomedical Engineering, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
- Department of Neuroscience, University of Rochester, 601 Elmwood Ave, Rochester, NY, 14642, USA
| |
Collapse
|
4
|
Drotos AC, Roberts MT. Identifying neuron types and circuit mechanisms in the auditory midbrain. Hear Res 2024; 442:108938. [PMID: 38141518 PMCID: PMC11000261 DOI: 10.1016/j.heares.2023.108938] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/08/2023] [Revised: 11/27/2023] [Accepted: 12/18/2023] [Indexed: 12/25/2023]
Abstract
The inferior colliculus (IC) is a critical computational hub in the central auditory pathway. From its position in the midbrain, the IC receives nearly all the ascending output from the lower auditory brainstem and provides the main source of auditory information to the thalamocortical system. In addition to being a crossroads for auditory circuits, the IC is rich with local circuits and contains more than five times as many neurons as the nuclei of the lower auditory brainstem combined. These results hint at the enormous computational power of the IC, and indeed, systems-level studies have identified numerous important transformations in sound coding that occur in the IC. However, despite decades of effort, the cellular mechanisms underlying IC computations and how these computations change following hearing loss have remained largely impenetrable. In this review, we argue that this challenge persists due to the surprisingly difficult problem of identifying the neuron types and circuit motifs that comprise the IC. After summarizing the extensive evidence pointing to a diversity of neuron types in the IC, we highlight the successes of recent efforts to parse this complexity using molecular markers to define neuron types. We conclude by arguing that the discovery of molecularly identifiable neuron types ushers in a new era for IC research marked by molecularly targeted recordings and manipulations. We propose that the ability to reproducibly investigate IC circuits at the neuronal level will lead to rapid advances in understanding the fundamental mechanisms driving IC computations and how these mechanisms shift following hearing loss.
Collapse
Affiliation(s)
- Audrey C Drotos
- Kresge Hearing Research Institute, Department of Otolaryngology - Head and Neck Surgery, University of Michigan, Ann Arbor, MI 48109, United States
| | - Michael T Roberts
- Kresge Hearing Research Institute, Department of Otolaryngology - Head and Neck Surgery, University of Michigan, Ann Arbor, MI 48109, United States; Department of Molecular and Integrative Physiology, University of Michigan, Ann Arbor, MI, 48109, United States.
| |
Collapse
|
5
|
Mitchell PW, Henry KS, Carney LH. Sensitivity to direction and velocity of fast frequency chirps in the inferior colliculus of awake rabbit. Hear Res 2023; 440:108915. [PMID: 37992517 PMCID: PMC10847965 DOI: 10.1016/j.heares.2023.108915] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/14/2023] [Revised: 10/20/2023] [Accepted: 11/14/2023] [Indexed: 11/24/2023]
Abstract
Neurons in the mammalian inferior colliculus (IC) are sensitive to the velocity (speed and direction) of fast frequency chirps contained in Schroeder-phase harmonic complexes (SCHR). However, IC neurons are also sensitive to stimulus periodicity, a prominent feature of SCHR stimuli. Here, to disentangle velocity sensitivity from periodicity tuning, we introduced a novel stimulus consisting of aperiodic random chirps. Extracellular, single-unit recordings were made in the IC of Dutch-belted rabbits in response to both SCHR and aperiodic chirps. Rate-velocity functions were constructed from aperiodic-chirp responses and compared to SCHR rate profiles, revealing interactions between stimulus periodicity and neural velocity sensitivity. A generalized linear model analysis demonstrated that periodicity tuning influences SCHR response rates more strongly than velocity sensitivity. Principal component analysis of rate-velocity functions revealed that neurons were more often sensitive to the direction of lower-velocity chirps and were less often sensitive to the direction of higher-velocity chirps. Overall, these results demonstrate that sensitivity to chirp velocity is common in the IC. Harmonic sounds with complex phase spectra, such as speech and music, contain chirps, and velocity sensitivity would shape IC responses to these sounds.
Collapse
Affiliation(s)
- Paul W Mitchell
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA
| | - Kenneth S Henry
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA; Departments of Neuroscience, University of Rochester, Rochester, NY, USA; Departments of Otolaryngology, University of Rochester, Rochester, NY, USA
| | - Laurel H Carney
- Department of Biomedical Engineering, University of Rochester, Rochester, NY, USA; Departments of Neuroscience, University of Rochester, Rochester, NY, USA.
| |
Collapse
|
6
|
Lindeberg T. A time-causal and time-recursive scale-covariant scale-space representation of temporal signals and past time. BIOLOGICAL CYBERNETICS 2023; 117:21-59. [PMID: 36689001 PMCID: PMC10160219 DOI: 10.1007/s00422-022-00953-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/18/2022] [Accepted: 11/21/2022] [Indexed: 05/05/2023]
Abstract
This article presents an overview of a theory for performing temporal smoothing on temporal signals in such a way that: (i) temporally smoothed signals at coarser temporal scales are guaranteed to constitute simplifications of corresponding temporally smoothed signals at any finer temporal scale (including the original signal) and (ii) the temporal smoothing process is both time-causal and time-recursive, in the sense that it does not require access to future information and can be performed with no other temporal memory buffer of the past than the resulting smoothed temporal scale-space representations themselves. For specific subsets of parameter settings for the classes of linear and shift-invariant temporal smoothing operators that obey this property, it is shown how temporal scale covariance can be additionally obtained, guaranteeing that if the temporal input signal is rescaled by a uniform temporal scaling factor, then also the resulting temporal scale-space representations of the rescaled temporal signal will constitute mere rescalings of the temporal scale-space representations of the original input signal, complemented by a shift along the temporal scale dimension. The resulting time-causal limit kernel that obeys this property constitutes a canonical temporal kernel for processing temporal signals in real-time scenarios when the regular Gaussian kernel cannot be used, because of its non-causal access to information from the future, and we cannot additionally require the temporal smoothing process to comprise a complementary memory of the past beyond the information contained in the temporal smoothing process itself, which in this way also serves as a multi-scale temporal memory of the past. We describe how the time-causal limit kernel relates to previously used temporal models, such as Koenderink's scale-time kernels and the ex-Gaussian kernel. We do also give an overview of how the time-causal limit kernel can be used for modelling the temporal processing in models for spatio-temporal and spectro-temporal receptive fields, and how it more generally has a high potential for modelling neural temporal response functions in a purely time-causal and time-recursive way, that can also handle phenomena at multiple temporal scales in a theoretically well-founded manner. We detail how this theory can be efficiently implemented for discrete data, in terms of a set of recursive filters coupled in cascade. Hence, the theory is generally applicable for both: (i) modelling continuous temporal phenomena over multiple temporal scales and (ii) digital processing of measured temporal signals in real time. We conclude by stating implications of the theory for modelling temporal phenomena in biological, perceptual, neural and memory processes by mathematical models, as well as implications regarding the philosophy of time and perceptual agents. Specifically, we propose that for A-type theories of time, as well as for perceptual agents, the notion of a non-infinitesimal inner temporal scale of the temporal receptive fields has to be included in representations of the present, where the inherent nonzero temporal delay of such time-causal receptive fields implies a need for incorporating predictions from the actual time-delayed present in the layers of a perceptual hierarchy, to make it possible for a representation of the perceptual present to constitute a representation of the environment with timing properties closer to the actual present.
Collapse
Affiliation(s)
- Tony Lindeberg
- Computational Brain Science Lab, Division of Computational Science and Technology, KTH Royal Institute of Technology, 100 44, Stockholm, Sweden.
| |
Collapse
|
7
|
Sadagopan S, Kar M, Parida S. Quantitative models of auditory cortical processing. Hear Res 2023; 429:108697. [PMID: 36696724 PMCID: PMC9928778 DOI: 10.1016/j.heares.2023.108697] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 12/17/2022] [Accepted: 01/12/2023] [Indexed: 01/15/2023]
Abstract
To generate insight from experimental data, it is critical to understand the inter-relationships between individual data points and place them in context within a structured framework. Quantitative modeling can provide the scaffolding for such an endeavor. Our main objective in this review is to provide a primer on the range of quantitative tools available to experimental auditory neuroscientists. Quantitative modeling is advantageous because it can provide a compact summary of observed data, make underlying assumptions explicit, and generate predictions for future experiments. Quantitative models may be developed to characterize or fit observed data, to test theories of how a task may be solved by neural circuits, to determine how observed biophysical details might contribute to measured activity patterns, or to predict how an experimental manipulation would affect neural activity. In complexity, quantitative models can range from those that are highly biophysically realistic and that include detailed simulations at the level of individual synapses, to those that use abstract and simplified neuron models to simulate entire networks. Here, we survey the landscape of recently developed models of auditory cortical processing, highlighting a small selection of models to demonstrate how they help generate insight into the mechanisms of auditory processing. We discuss examples ranging from models that use details of synaptic properties to explain the temporal pattern of cortical responses to those that use modern deep neural networks to gain insight into human fMRI data. We conclude by discussing a biologically realistic and interpretable model that our laboratory has developed to explore aspects of vocalization categorization in the auditory pathway.
Collapse
Affiliation(s)
- Srivatsun Sadagopan
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA; Department of Bioengineering, University of Pittsburgh, Pittsburgh, PA, USA; Department of Communication Science and Disorders, University of Pittsburgh, Pittsburgh, PA, USA.
| | - Manaswini Kar
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA; Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, PA, USA
| | - Satyabrata Parida
- Department of Neurobiology, University of Pittsburgh, Pittsburgh, PA, USA; Center for Neuroscience, University of Pittsburgh, Pittsburgh, PA, USA
| |
Collapse
|
8
|
He F, Stevenson IH, Escabí MA. Two stages of bandwidth scaling drives efficient neural coding of natural sounds. PLoS Comput Biol 2023; 19:e1010862. [PMID: 36787338 PMCID: PMC9970106 DOI: 10.1371/journal.pcbi.1010862] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2022] [Revised: 02/27/2023] [Accepted: 01/09/2023] [Indexed: 02/15/2023] Open
Abstract
Theories of efficient coding propose that the auditory system is optimized for the statistical structure of natural sounds, yet the transformations underlying optimal acoustic representations are not well understood. Using a database of natural sounds including human speech and a physiologically-inspired auditory model, we explore the consequences of peripheral (cochlear) and mid-level (auditory midbrain) filter tuning transformations on the representation of natural sound spectra and modulation statistics. Whereas Fourier-based sound decompositions have constant time-frequency resolution at all frequencies, cochlear and auditory midbrain filters bandwidths increase proportional to the filter center frequency. This form of bandwidth scaling produces a systematic decrease in spectral resolution and increase in temporal resolution with increasing frequency. Here we demonstrate that cochlear bandwidth scaling produces a frequency-dependent gain that counteracts the tendency of natural sound power to decrease with frequency, resulting in a whitened output representation. Similarly, bandwidth scaling in mid-level auditory filters further enhances the representation of natural sounds by producing a whitened modulation power spectrum (MPS) with higher modulation entropy than both the cochlear outputs and the conventional Fourier MPS. These findings suggest that the tuning characteristics of the peripheral and mid-level auditory system together produce a whitened output representation in three dimensions (frequency, temporal and spectral modulation) that reduces redundancies and allows for a more efficient use of neural resources. This hierarchical multi-stage tuning strategy is thus likely optimized to extract available information and may underlies perceptual sensitivity to natural sounds.
Collapse
Affiliation(s)
- Fengrong He
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Ian H. Stevenson
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- The Connecticut Institute for Brain and Cognitive Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- The Connecticut Institute for Brain and Cognitive Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- * E-mail:
| |
Collapse
|
9
|
Macias S, Bakshi K, Troyer T, Smotherman M. The prefrontal cortex of the Mexican free-tailed bat is more selective to communication calls than primary auditory cortex. J Neurophysiol 2022; 128:634-648. [PMID: 35975923 PMCID: PMC9448334 DOI: 10.1152/jn.00436.2021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Revised: 07/20/2022] [Accepted: 08/05/2022] [Indexed: 11/22/2022] Open
Abstract
In this study, we examined the auditory responses of a prefrontal area, the frontal auditory field (FAF), of an echolocating bat (Tadarida brasiliensis) and presented a comparative analysis of the neuronal response properties between the FAF and the primary auditory cortex (A1). We compared single-unit responses from the A1 and the FAF elicited by pure tones, downward frequency-modulated sweeps (dFMs), and species-specific vocalizations. Unlike the A1, FAFs were not frequency tuned. However, progressive increases in dFM sweep rate elicited a systematic increase of response precision, a phenomenon that does not take place in the A1. Call selectivity was higher in the FAF versus A1. We calculated the neuronal spectrotemporal receptive fields (STRFs) and spike-triggered averages (STAs) to predict responses to the communication calls and provide an explanation for the differences in call selectivity between the FAF and A1. In the A1, we found a high correlation between predicted and evoked responses. However, we did not generate reasonable STRFs in the FAF, and the prediction based on the STAs showed lower correlation coefficient than that of the A1. This suggests nonlinear response properties in the FAF that are stronger than the linear response properties in the A1. Stimulating with a call sequence increased call selectivity in the A1, but it remained unchanged in the FAF. These data are consistent with a role for the FAF in assessing distinctive acoustic features downstream of A1, similar to the role proposed for primate ventrolateral prefrontal cortex.NEW & NOTEWORTHY In this study, we examined the neuronal responses of a frontal cortical area in an echolocating bat to behaviorally relevant acoustic stimuli and compared them with those in the primary auditory cortex (A1). In contrast to the A1, neurons in the bat frontal auditory field are not frequency tuned but showed a higher selectivity for social signals such as communication calls. The results presented here indicate that the frontal auditory field may represent an additional processing center for behaviorally relevant sounds.
Collapse
Affiliation(s)
- Silvio Macias
- Department of Biology, Texas A&M University, College Station, Texas
| | - Kushal Bakshi
- Institute for Neuroscience, Texas A&M University, College Station, Texas
| | - Todd Troyer
- Department of Neuroscience, Developmental and Regenerative Biology, University of Texas at San Antonio, San Antonio, Texas
| | - Michael Smotherman
- Department of Biology, Texas A&M University, College Station, Texas
- Institute for Neuroscience, Texas A&M University, College Station, Texas
| |
Collapse
|
10
|
Tuned in to communication sounds: Neuronal sensitivity in the túngara frog midbrain to frequency modulated signals. PLoS One 2022; 17:e0268383. [PMID: 35587486 PMCID: PMC9119527 DOI: 10.1371/journal.pone.0268383] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2021] [Accepted: 04/28/2022] [Indexed: 11/19/2022] Open
Abstract
For complex communication signals, it is often difficult to identify the information-bearing elements and their parameters necessary to elicit functional behavior. Consequently, it may be difficult to design stimuli that test how neurons contribute to communicative processing. For túngara frogs (Physalaemus pustulosus), however, previous behavioral testing with numerous stimuli showed that a particular frequency modulated (FM) transition in the male call is required to elicit phonotaxis and vocal responses. Modeled on such behavioral experiments, we used awake in vivo recordings of single units in the midbrain to determine if their excitation was biased to behaviorally important FM parameters. Comparisons of stimulus driven action potentials revealed greatest excitation to the behaviorally important FM transition: a downward FM sweep or step that crosses ~600 Hz. Previous studies using long-duration acoustic exposure found immediate early gene expression in many midbrain neurons to be most sensitive to similar FM. However, those data could not determine if FM coding was accomplished by the population and/or individual neurons. Our data suggest both coding schemes could operate, as 1) individual neurons are more sensitive to the behaviorally significant FM transition and 2) when single unit recordings are analytically combined across cells, the combined code can produce high stimulus discrimination (FM vs. noise driven excitation), approaching that found in behavioral discrimination of call vs. noise.
Collapse
|
11
|
Gentile Polese A, Nigam S, Hurley LM. 5-HT1A Receptors Alter Temporal Responses to Broadband Vocalizations in the Mouse Inferior Colliculus Through Response Suppression. Front Neural Circuits 2021; 15:718348. [PMID: 34512276 PMCID: PMC8430226 DOI: 10.3389/fncir.2021.718348] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2021] [Accepted: 07/19/2021] [Indexed: 01/21/2023] Open
Abstract
Neuromodulatory systems may provide information on social context to auditory brain regions, but relatively few studies have assessed the effects of neuromodulation on auditory responses to acoustic social signals. To address this issue, we measured the influence of the serotonergic system on the responses of neurons in a mouse auditory midbrain nucleus, the inferior colliculus (IC), to vocal signals. Broadband vocalizations (BBVs) are human-audible signals produced by mice in distress as well as by female mice in opposite-sex interactions. The production of BBVs is context-dependent in that they are produced both at early stages of interactions as females physically reject males and at later stages as males mount females. Serotonin in the IC of males corresponds to these events, and is elevated more in males that experience less female rejection. We measured the responses of single IC neurons to five recorded examples of BBVs in anesthetized mice. We then locally activated the 5-HT1A receptor through iontophoretic application of 8-OH-DPAT. IC neurons showed little selectivity for different BBVs, but spike trains were characterized by local regions of high spike probability, which we called "response features." Response features varied across neurons and also across calls for individual neurons, ranging from 1 to 7 response features for responses of single neurons to single calls. 8-OH-DPAT suppressed spikes and also reduced the numbers of response features. The weakest response features were the most likely to disappear, suggestive of an "iceberg"-like effect in which activation of the 5-HT1A receptor suppressed weakly suprathreshold response features below the spiking threshold. Because serotonin in the IC is more likely to be elevated for mounting-associated BBVs than for rejection-associated BBVs, these effects of the 5-HT1A receptor could contribute to the differential auditory processing of BBVs in different behavioral subcontexts.
Collapse
Affiliation(s)
- Arianna Gentile Polese
- Department of Cell and Developmental Biology, University of Colorado Anschutz Medical Campus, Aurora, CO, United States
- Department of Biology, Program in Neuroscience, Indiana University Bloomington, Bloomington, IN, United States
| | - Sunny Nigam
- Department of Neurobiology and Anatomy, McGovern Medical School, The University of Texas Health Science Center at Houston, Houston, TX, United States
- Department of Physics, Indiana University Bloomington, Bloomington, IN, United States
| | - Laura M. Hurley
- Department of Neurobiology and Anatomy, McGovern Medical School, The University of Texas Health Science Center at Houston, Houston, TX, United States
| |
Collapse
|
12
|
Chitradurga Achutha A, Peremans H, Firzlaff U, Vanderelst D. Efficient encoding of spectrotemporal information for bat echolocation. PLoS Comput Biol 2021; 17:e1009052. [PMID: 34181643 PMCID: PMC8270447 DOI: 10.1371/journal.pcbi.1009052] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2020] [Revised: 07/09/2021] [Accepted: 05/07/2021] [Indexed: 12/04/2022] Open
Abstract
In most animals, natural stimuli are characterized by a high degree of redundancy, limiting the ensemble of ecologically valid stimuli to a significantly reduced subspace of the representation space. Neural encodings can exploit this redundancy and increase sensing efficiency by generating low-dimensional representations that retain all information essential to support behavior. In this study, we investigate whether such an efficient encoding can be found to support a broad range of echolocation tasks in bats. Starting from an ensemble of echo signals collected with a biomimetic sonar system in natural indoor and outdoor environments, we use independent component analysis to derive a low-dimensional encoding of the output of a cochlear model. We show that this compressive encoding retains all essential information. To this end, we simulate a range of psycho-acoustic experiments with bats. In these simulations, we train a set of neural networks to use the encoded echoes as input while performing the experiments. The results show that the neural networks’ performance is at least as good as that of the bats. We conclude that our results indicate that efficient encoding of echo information is feasible and, given its many advantages, very likely to be employed by bats. Previous studies have demonstrated that low-dimensional encodings allow for task resolution at a relatively high level. In contrast to previous work in this area, we show that high performance can also be achieved when low-dimensional filters are derived from a data set of realistic echo signals, not tailored to specific experimental conditions. We show that complex (and simple) echoes from real environments can be efficiently and effectively represented using a small set of filters. Critically, we show that high performance across a range of tasks can be achieved when low-dimensional filters are derived from a data set of realistic echo signals, not tailored to specific experimental conditions. The redundancy in echoic information opens up the opportunity for efficient encoding, reducing the computational load of echo processing as well as the memory load for storing the information. Therefore, we predict the auditory system of bats to capitalize on this opportunity for efficient coding by implementing filters with spectrotemporal properties akin to those hypothesized here. Indeed, the filters we obtain here are similar to those found in other animals and other sensing capabilities. Our results indicate that bats could exploit the redundancy in sonar signals to implement an efficient neural encoding of the relevant information.
Collapse
Affiliation(s)
- Adarsh Chitradurga Achutha
- Mechanical and Materials Engineering, University of Cincinnati, Cincinnati, Ohio, United States of America
| | - Herbert Peremans
- Department of Engineering Management, University of Antwerp, Antwerp, Belgium
| | - Uwe Firzlaff
- Chair of Zoology, School of Life Sciences, Technical University of Munich, Freising, Germany
| | - Dieter Vanderelst
- Department of Biological Sciences, University of Cincinnati, Cincinnati, Ohio, United States of America
- * E-mail:
| |
Collapse
|
13
|
Natural Statistics as Inference Principles of Auditory Tuning in Biological and Artificial Midbrain Networks. eNeuro 2021; 8:ENEURO.0525-20.2021. [PMID: 33947687 PMCID: PMC8211468 DOI: 10.1523/eneuro.0525-20.2021] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2020] [Revised: 03/10/2021] [Accepted: 04/27/2021] [Indexed: 12/04/2022] Open
Abstract
Bats provide a powerful mammalian model to explore the neural representation of complex sounds, as they rely on hearing to survive in their environment. The inferior colliculus (IC) is a central hub of the auditory system that receives converging projections from the ascending pathway and descending inputs from auditory cortex. In this work, we build an artificial neural network to replicate auditory characteristics in IC neurons of the big brown bat. We first test the hypothesis that spectro-temporal tuning of IC neurons is optimized to represent the natural statistics of conspecific vocalizations. We estimate spectro-temporal receptive fields (STRFs) of IC neurons and compare tuning characteristics to statistics of bat calls. The results indicate that the FM tuning of IC neurons is matched with the statistics. Then, we investigate this hypothesis on the network optimized to represent natural sound statistics and to compare its output with biological responses. We also estimate biomimetic STRFs from the artificial network and correlate their characteristics to those of biological neurons. Tuning properties of both biological and artificial neurons reveal strong agreement along both spectral and temporal dimensions, and suggest the presence of nonlinearity, sparsity, and complexity constraints that underlie the neural representation in the auditory midbrain. Additionally, the artificial neurons replicate IC neural activities in discrimination of social calls, and provide simulated results for a noise robust discrimination. In this way, the biomimetic network allows us to infer the neural mechanisms by which the bat’s IC processes natural sounds used to construct the auditory scene.
Collapse
|
14
|
Aponte DA, Handy G, Kline AM, Tsukano H, Doiron B, Kato HK. Recurrent network dynamics shape direction selectivity in primary auditory cortex. Nat Commun 2021; 12:314. [PMID: 33436635 PMCID: PMC7804939 DOI: 10.1038/s41467-020-20590-6] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2020] [Accepted: 12/11/2020] [Indexed: 02/03/2023] Open
Abstract
Detecting the direction of frequency modulation (FM) is essential for vocal communication in both animals and humans. Direction-selective firing of neurons in the primary auditory cortex (A1) has been classically attributed to temporal offsets between feedforward excitatory and inhibitory inputs. However, it remains unclear how cortical recurrent circuitry contributes to this computation. Here, we used two-photon calcium imaging and whole-cell recordings in awake mice to demonstrate that direction selectivity is not caused by temporal offsets between synaptic currents, but by an asymmetry in total synaptic charge between preferred and non-preferred directions. Inactivation of cortical somatostatin-expressing interneurons (SOM cells) reduced direction selectivity, revealing its cortical contribution. Our theoretical models showed that charge asymmetry arises due to broad spatial topography of SOM cell-mediated inhibition which regulates signal amplification in strongly recurrent circuitry. Together, our findings reveal a major contribution of recurrent network dynamics in shaping cortical tuning to behaviorally relevant complex sounds.
Collapse
Affiliation(s)
- Destinee A Aponte
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
- Neuroscience Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
| | - Gregory Handy
- Departments of Neurobiology and Statistics, University of Chicago, Chicago, IL, USA
- Department of Mathematics, University of Pittsburgh, Pittsburgh, USA
- Grossman Center for Quantitative Biology and Human Behavior, University of Chicago, Chicago, IL, USA
| | - Amber M Kline
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
- Neuroscience Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
| | - Hiroaki Tsukano
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
- Neuroscience Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
| | - Brent Doiron
- Departments of Neurobiology and Statistics, University of Chicago, Chicago, IL, USA
- Department of Mathematics, University of Pittsburgh, Pittsburgh, USA
- Grossman Center for Quantitative Biology and Human Behavior, University of Chicago, Chicago, IL, USA
| | - Hiroyuki K Kato
- Department of Psychiatry, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA.
- Neuroscience Center, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA.
- Carolina Institute for Developmental Disabilities, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA.
| |
Collapse
|
15
|
Logerot P, Smith PF, Wild M, Kubke MF. Auditory processing in the zebra finch midbrain: single unit responses and effect of rearing experience. PeerJ 2020; 8:e9363. [PMID: 32775046 PMCID: PMC7384439 DOI: 10.7717/peerj.9363] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2019] [Accepted: 05/26/2020] [Indexed: 11/26/2022] Open
Abstract
In birds the auditory system plays a key role in providing the sensory input used to discriminate between conspecific and heterospecific vocal signals. In those species that are known to learn their vocalizations, for example, songbirds, it is generally considered that this ability arises and is manifest in the forebrain, although there is no a priori reason why brainstem components of the auditory system could not also play an important part. To test this assumption, we used groups of normal reared and cross-fostered zebra finches that had previously been shown in behavioural experiments to reduce their preference for conspecific songs subsequent to cross fostering experience with Bengalese finches, a related species with a distinctly different song. The question we asked, therefore, is whether this experiential change also changes the bias in favour of conspecific song displayed by auditory midbrain units of normally raised zebra finches. By recording the responses of single units in MLd to a variety of zebra finch and Bengalese finch songs in both normally reared and cross-fostered zebra finches, we provide a positive answer to this question. That is, the difference in response to conspecific and heterospecific songs seen in normal reared zebra finches is reduced following cross-fostering. In birds the virtual absence of mammalian-like cortical projections upon auditory brainstem nuclei argues against the interpretation that MLd units change, as observed in the present experiments, as a result of top-down influences on sensory processing. Instead, it appears that MLd units can be influenced significantly by sensory inputs arising directly from a change in auditory experience during development.
Collapse
Affiliation(s)
- Priscilla Logerot
- Anatomy and Medical Imaging, University of Auckland, University of Auckland, Auckland, New Zealand
| | - Paul F. Smith
- Dept. of Pharmacology and Toxicology, School of Biomedical Sciences, Brain Health Research Centre, Brain Research New Zealand, and Eisdell Moore Centre, University of Otago, Dunedin, New Zealand
| | - Martin Wild
- Anatomy and Medical Imaging and Eisdell Moore Centre, University of Auckland, University of Auckland, Auckland, New Zealand
| | - M. Fabiana Kubke
- Anatomy and Medical Imaging, Centre for Brain Research and Eisdell Moore Centre, University of Auckland, University of Auckland, Auckland, New Zealand
| |
Collapse
|
16
|
Macias S, Bakshi K, Smotherman M. Functional organization of the primary auditory cortex of the free-tailed bat Tadarida brasiliensis. J Comp Physiol A Neuroethol Sens Neural Behav Physiol 2020; 206:429-440. [PMID: 32036404 DOI: 10.1007/s00359-020-01406-w] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2019] [Revised: 01/20/2020] [Accepted: 01/22/2020] [Indexed: 12/19/2022]
Abstract
The Mexican free-tailed bat, Tadarida brasiliensis, is a fast-flying bat that hunts by biosonar at high altitudes in open space. The auditory periphery and ascending auditory pathways have been described in great detail for this species, but nothing is yet known about its auditory cortex. Here we describe the topographical organization of response properties in the primary auditory cortex (AC) of the Mexican free-tailed bat with emphasis on the sensitivity for FM sweeps and echo-delay tuning. Responses of 716 units to pure tones and of 373 units to FM sweeps and FM-FM pairs were recorded extracellularly using multielectrode arrays in anesthetized bats. A general tonotopy was confirmed with low frequencies represented caudally and high frequencies represented rostrally. Characteristic frequencies (CF) ranged from 15 to 70 kHz, and fifty percent of CFs fell between 20 and 30 kHz, reflecting a hyper-representation of a bandwidth corresponding to search-phase echolocation pulses. Most units showed a stronger response to downward rather than upward FM sweeps and forty percent of the neurons interspersed throughout AC (150/371) showed echo-delay sensitivity to FM-FM pairs. Overall, the results illustrate that the free-tailed bat auditory cortex is organized similarly to that of other FM-type insectivorous bats.
Collapse
Affiliation(s)
- Silvio Macias
- Department of Biology, Texas A&M University, College Station, TX, 77843, USA.
| | - Kushal Bakshi
- Department of Biology, Texas A&M University, College Station, TX, 77843, USA
| | - Michael Smotherman
- Department of Biology, Texas A&M University, College Station, TX, 77843, USA
| |
Collapse
|
17
|
Gourévitch B, Mahrt EJ, Bakay W, Elde C, Portfors CV. GABA A receptors contribute more to rate than temporal coding in the IC of awake mice. J Neurophysiol 2020; 123:134-148. [PMID: 31721644 DOI: 10.1152/jn.00377.2019] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/11/2022] Open
Abstract
Speech is our most important form of communication, yet we have a poor understanding of how communication sounds are processed by the brain. Mice make great model organisms to study neural processing of communication sounds because of their rich repertoire of social vocalizations and because they have brain structures analogous to humans, such as the auditory midbrain nucleus inferior colliculus (IC). Although the combined roles of GABAergic and glycinergic inhibition on vocalization selectivity in the IC have been studied to a limited degree, the discrete contributions of GABAergic inhibition have only rarely been examined. In this study, we examined how GABAergic inhibition contributes to shaping responses to pure tones as well as selectivity to complex sounds in the IC of awake mice. In our set of long-latency neurons, we found that GABAergic inhibition extends the evoked firing rate range of IC neurons by lowering the baseline firing rate but maintaining the highest probability of firing rate. GABAergic inhibition also prevented IC neurons from bursting in a spontaneous state. Finally, we found that although GABAergic inhibition shaped the spectrotemporal response to vocalizations in a nonlinear fashion, it did not affect the neural code needed to discriminate vocalizations, based either on spiking patterns or on firing rate. Overall, our results emphasize that even if GABAergic inhibition generally decreases the firing rate, it does so while maintaining or extending the abilities of neurons in the IC to code the wide variety of sounds that mammals are exposed to in their daily lives.NEW & NOTEWORTHY GABAergic inhibition adds nonlinearity to neuronal response curves. This increases the neuronal range of evoked firing rate by reducing baseline firing. GABAergic inhibition prevents bursting responses from neurons in a spontaneous state, reducing noise in the temporal coding of the neuron. This could result in improved signal transmission to the cortex.
Collapse
Affiliation(s)
- Boris Gourévitch
- Institut de l'Audition, Institut Pasteur, INSERM, Sorbonne Université, F-75012 Paris, France.,CNRS, France
| | - Elena J Mahrt
- School of Biological Sciences, Washington State University, Vancouver, Washington
| | - Warren Bakay
- Institut de l'Audition, Institut Pasteur, INSERM, Sorbonne Université, F-75012 Paris, France
| | - Cameron Elde
- School of Biological Sciences, Washington State University, Vancouver, Washington
| | - Christine V Portfors
- School of Biological Sciences, Washington State University, Vancouver, Washington
| |
Collapse
|
18
|
Crommett LE, Madala D, Yau JM. Multisensory perceptual interactions between higher-order temporal frequency signals. J Exp Psychol Gen 2019; 148:1124-1137. [PMID: 30335446 PMCID: PMC6472995 DOI: 10.1037/xge0000513] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Naturally occurring signals in audition and touch can be complex and marked by temporal variations in frequency and amplitude. Auditory frequency sweep processing has been studied extensively; however, much less is known about sweep processing in touch because studies have primarily focused on the perception of simple sinusoidal vibrations. Given the extensive interactions between audition and touch in the frequency processing of pure tone signals, we reasoned that these senses might also interact in the processing of higher-order frequency representations like sweeps. In a series of psychophysical experiments, we characterized the influence of auditory distractors on the ability of participants to discriminate tactile frequency sweeps. Auditory frequency sweeps systematically biased the tactile perception of sweep direction. Importantly, auditory cues exerted little influence on tactile sweep direction perception when the sounds and vibrations occupied different absolute frequency ranges or when the sounds consisted of intensity sweeps. Thus, audition and touch interact in frequency sweep perception in a frequency- and feature-specific manner. Our results demonstrate that audio-tactile interactions are not constrained to the processing of simple sinusoids. Because higher-order frequency representations may be synthesized from simpler representations, our findings imply that multisensory interactions in the temporal frequency domain span multiple hierarchical levels in sensory processing. (PsycINFO Database Record (c) 2019 APA, all rights reserved).
Collapse
Affiliation(s)
- Lexi E. Crommett
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas 77030, USA
| | | | - Jeffrey M. Yau
- Department of Neuroscience, Baylor College of Medicine, Houston, Texas 77030, USA
| |
Collapse
|
19
|
Chen C, Read HL, Escabí MA. A temporal integration mechanism enhances frequency selectivity of broadband inputs to inferior colliculus. PLoS Biol 2019; 17:e2005861. [PMID: 31233489 PMCID: PMC6611646 DOI: 10.1371/journal.pbio.2005861] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 07/05/2019] [Accepted: 05/22/2019] [Indexed: 11/18/2022] Open
Abstract
Accurately resolving frequency components in sounds is essential for sound recognition, yet there is little direct evidence for how frequency selectivity is preserved or newly created across auditory structures. We demonstrate that prepotentials (PPs) with physiological properties resembling presynaptic potentials from broadly tuned brainstem inputs can be recorded concurrently with postsynaptic action potentials in inferior colliculus (IC). These putative brainstem inputs (PBIs) are broadly tuned and exhibit delayed and spectrally interleaved excitation and inhibition not present in the simultaneously recorded IC neurons (ICNs). A sharpening of tuning is accomplished locally at the expense of spike-timing precision through nonlinear temporal integration of broadband inputs. A neuron model replicates the finding and demonstrates that temporal integration alone can degrade timing precision while enhancing frequency tuning through interference of spectrally in- and out-of-phase inputs. These findings suggest that, in contrast to current models that require local inhibition, frequency selectivity can be sharpened through temporal integration, thus supporting an alternative computational strategy to quickly refine frequency selectivity.
Collapse
Affiliation(s)
- Chen Chen
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Heather L. Read
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| |
Collapse
|
20
|
Hörpel SG, Firzlaff U. Processing of fast amplitude modulations in bat auditory cortex matches communication call-specific sound features. J Neurophysiol 2019; 121:1501-1512. [PMID: 30785811 DOI: 10.1152/jn.00748.2018] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Bats use a large repertoire of calls for social communication. In the bat Phyllostomus discolor, social communication calls are often characterized by sinusoidal amplitude and frequency modulations with modulation frequencies in the range of 100-130 Hz. However, peaks in mammalian auditory cortical modulation transfer functions are typically limited to modulation frequencies below 100 Hz. We investigated the coding of sinusoidally amplitude modulated sounds in auditory cortical neurons in P. discolor by constructing rate and temporal modulation transfer functions. Neuronal responses to playbacks of various communication calls were additionally recorded and compared with the neurons' responses to sinusoidally amplitude-modulated sounds. Cortical neurons in the posterior dorsal field of the auditory cortex were tuned to unusually high modulation frequencies: rate modulation transfer functions often peaked around 130 Hz (median: 87 Hz), and the median of the highest modulation frequency that evoked significant phase-locking was also 130 Hz. Both values are much higher than reported from the auditory cortex of other mammals, with more than 51% of the units preferring modulation frequencies exceeding 100 Hz. Conspicuously, the fast modulations preferred by the neurons match the fast amplitude and frequency modulations of prosocial, and mostly of aggressive, communication calls in P. discolor. We suggest that the preference for fast amplitude modulations in the P. discolor dorsal auditory cortex serves to reliably encode the fast modulations seen in their communication calls. NEW & NOTEWORTHY Neural processing of temporal sound features is crucial for the analysis of communication calls. In bats, these calls are often characterized by fast temporal envelope modulations. Because auditory cortex neurons typically encode only low modulation frequencies, it is unclear how species-specific vocalizations are cortically processed. We show that auditory cortex neurons in the bat Phyllostomus discolor encode fast temporal envelope modulations. This property improves response specificity to communication calls and thus might support species-specific communication.
Collapse
Affiliation(s)
- Stephen Gareth Hörpel
- Chair of Zoology, Department of Animal Sciences, Technical University of Munich , Freising , Germany
| | - Uwe Firzlaff
- Chair of Zoology, Department of Animal Sciences, Technical University of Munich , Freising , Germany
| |
Collapse
|
21
|
Zhang Q, Hu X, Hong B, Zhang B. A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex. PLoS Comput Biol 2019; 15:e1006766. [PMID: 30742609 PMCID: PMC6386396 DOI: 10.1371/journal.pcbi.1006766] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2018] [Revised: 02/22/2019] [Accepted: 12/21/2018] [Indexed: 12/03/2022] Open
Abstract
The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway. When speech enters the ear, it is subjected to a series of processing stages prior to arriving at the auditory cortex. Neurons acting at different processing stages have different response properties. For example, at the auditory midbrain, a neuron may specifically detect the onsets of a frequency component in the speech, whereas in the auditory cortex, a neuron may specifically detect phonetic features. The encoding mechanisms underlying these neuronal functions remain unclear. To address this issue, we designed a hierarchical sparse coding model, inspired by the sparse activity of neurons in the sensory system, to learn features in speech signals. We found that the computing units in different layers exhibited hierarchical extraction of speech sound features, similar to those of neurons in the auditory midbrain and auditory cortex, although the computational principles in these layers were the same. The results suggest that sparse coding and max pooling represent universal computational principles throughout the auditory pathway.
Collapse
Affiliation(s)
- Qingtian Zhang
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
| | - Xiaolin Hu
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
- Center for Brain-Inspired Computing Research (CBICR), Tsinghua University, Beijing, China
- * E-mail:
| | - Bo Hong
- School of Medicine, Tsinghua University, Beijing, China
| | - Bo Zhang
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
- Center for Brain-Inspired Computing Research (CBICR), Tsinghua University, Beijing, China
| |
Collapse
|
22
|
Teng X, Tian X, Doelling K, Poeppel D. Theta band oscillations reflect more than entrainment: behavioral and neural evidence demonstrates an active chunking process. Eur J Neurosci 2018; 48:2770-2782. [PMID: 29044763 PMCID: PMC5904023 DOI: 10.1111/ejn.13742] [Citation(s) in RCA: 33] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2017] [Revised: 08/16/2017] [Accepted: 09/28/2017] [Indexed: 11/29/2022]
Abstract
Parsing continuous acoustic streams into perceptual units is fundamental to auditory perception. Previous studies have uncovered a cortical entrainment mechanism in the delta and theta bands (~1-8 Hz) that correlates with formation of perceptual units in speech, music, and other quasi-rhythmic stimuli. Whether cortical oscillations in the delta-theta bands are passively entrained by regular acoustic patterns or play an active role in parsing the acoustic stream is debated. Here, we investigate cortical oscillations using novel stimuli with 1/f modulation spectra. These 1/f signals have no rhythmic structure but contain information over many timescales because of their broadband modulation characteristics. We chose 1/f modulation spectra with varying exponents of f, which simulate the dynamics of environmental noise, speech, vocalizations, and music. While undergoing magnetoencephalography (MEG) recording, participants listened to 1/f stimuli and detected embedded target tones. Tone detection performance varied across stimuli of different exponents and can be explained by local signal-to-noise ratio computed using a temporal window around 200 ms. Furthermore, theta band oscillations, surprisingly, were observed for all stimuli, but robust phase coherence was preferentially displayed by stimuli with exponents 1 and 1.5. We constructed an auditory processing model to quantify acoustic information on various timescales and correlated the model outputs with the neural results. We show that cortical oscillations reflect a chunking of segments, > 200 ms. These results suggest an active auditory segmentation mechanism, complementary to entrainment, operating on a timescale of ~200 ms to organize acoustic information.
Collapse
Affiliation(s)
| | - Xing Tian
- New York University Shanghai, Shanghai, China, 200122
- NYU-ECNU Institute of Brain and Cognitive Science at NYU Shanghai, Shanghai, China 200062
| | - Keith Doelling
- Department of Psychology, New York University, New York, NY, USA 10003
- Center for Neural Science, New York University, New York, NY, USA 10003
| | - David Poeppel
- Max-Planck-Institute, 60322 Frankfurt, Germany
- Department of Psychology, New York University, New York, NY, USA 10003
| |
Collapse
|
23
|
Jones TK, Wohlgemuth MJ, Conner WE. Active acoustic interference elicits echolocation changes in heterospecific bats. ACTA ACUST UNITED AC 2018; 221:jeb.176511. [PMID: 29950451 DOI: 10.1242/jeb.176511] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2017] [Accepted: 06/14/2018] [Indexed: 11/20/2022]
Abstract
Echolocating bats often forage in the presence of both conspecific and heterospecific individuals, which have the potential to produce acoustic interference. Recent studies have shown that at least one bat species, the Brazilian free-tailed bat (Tadarida brasiliensis), produces specialized social signals that disrupt the sonar of conspecific competitors. We herein discuss the differences between passive and active jamming signals and test whether heterospecific jamming occurs in species overlapping spatiotemporally, as well as whether such interference elicits a jamming avoidance response. We compare the capture rates of tethered moths and the echolocation parameters of big brown bats (Eptesicus fuscus) challenged with the playback of the jamming signal normally produced by Brazilian free-tailed bats and playback of deconstructed versions of this signal. There were no differences in the capture rates of targets with and without the jamming signal, although significant changes in both spectral and temporal features of the bats' echolocation were observed. These changes are consistent with improvements of the signal-to-noise ratio in the presence of acoustic interference. Accordingly, we propose to expand the traditional definition of the jamming avoidance response, stating that echolocation changes in response to interference should decrease similarity between the two signals, to include any change that increases the ability to separate returning echoes from active jamming stimuli originating from conspecific and heterospecific organisms. Flexibility in echolocation is an important characteristic for overcoming various forms of acoustic interference and may serve a purpose in interspecific interactions as well as intraspecific ones.
Collapse
Affiliation(s)
- Te K Jones
- Department of Biology, Wake Forest University, Winston-Salem, NC 27109, USA
| | - Melville J Wohlgemuth
- Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, MD 21218, USA
| | - William E Conner
- Department of Biology, Wake Forest University, Winston-Salem, NC 27109, USA
| |
Collapse
|
24
|
Peng F, Innes-Brown H, McKay CM, Fallon JB, Zhou Y, Wang X, Hu N, Hou W. Temporal Coding of Voice Pitch Contours in Mandarin Tones. Front Neural Circuits 2018; 12:55. [PMID: 30087597 PMCID: PMC6066958 DOI: 10.3389/fncir.2018.00055] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2017] [Accepted: 06/27/2018] [Indexed: 11/13/2022] Open
Abstract
Accurate perception of time-variant pitch is important for speech recognition, particularly for tonal languages with different lexical tones such as Mandarin, in which different tones convey different semantic information. Previous studies reported that the auditory nerve and cochlear nucleus can encode different pitches through phase-locked neural activities. However, little is known about how the inferior colliculus (IC) encodes the time-variant periodicity pitch of natural speech. In this study, the Mandarin syllable /ba/ pronounced with four lexical tones (flat, rising, falling then rising and falling) were used as stimuli. Local field potentials (LFPs) and single neuron activity were simultaneously recorded from 90 sites within contralateral IC of six urethane-anesthetized and decerebrate guinea pigs in response to the four stimuli. Analysis of the temporal information of LFPs showed that 93% of the LFPs exhibited robust encoding of periodicity pitch. Pitch strength of LFPs derived from the autocorrelogram was significantly (p < 0.001) stronger for rising tones than flat and falling tones. Pitch strength are also significantly increased (p < 0.05) with the characteristic frequency (CF). On the other hand, only 47% (42 or 90) of single neuron activities were significantly synchronized to the fundamental frequency of the stimulus suggesting that the temporal spiking pattern of single IC neuron could encode the time variant periodicity pitch of speech robustly. The difference between the number of LFPs and single neurons that encode the time-variant F0 voice pitch supports the notion of a transition at the level of IC from direct temporal coding in the spike trains of individual neurons to other form of neural representation.
Collapse
Affiliation(s)
- Fei Peng
- Key Laboratory of Biorheological Science and Technology of Ministry of Education, Bioengineering College of Chongqing University, Chongqing, China
- Collaborative Innovation Center for Brain Science, Chongqing University, Chongqing, China
| | - Hamish Innes-Brown
- Bionics Institute, East Melbourne, VIC, Australia
- Department of Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia
| | - Colette M. McKay
- Bionics Institute, East Melbourne, VIC, Australia
- Department of Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia
| | - James B. Fallon
- Bionics Institute, East Melbourne, VIC, Australia
- Department of Medical Bionics Department, University of Melbourne, Melbourne, VIC, Australia
- Department of Otolaryngology, University of Melbourne, Melbourne, VIC, Australia
| | - Yi Zhou
- Chongqing Key Laboratory of Neurobiology, Department of Neurobiology, Third Military Medical University, Chongqing, China
| | - Xing Wang
- Key Laboratory of Biorheological Science and Technology of Ministry of Education, Bioengineering College of Chongqing University, Chongqing, China
- Chongqing Medical Electronics Engineering Technology Research Center, Chongqing University, Chongqing, China
| | - Ning Hu
- Key Laboratory of Biorheological Science and Technology of Ministry of Education, Bioengineering College of Chongqing University, Chongqing, China
- Collaborative Innovation Center for Brain Science, Chongqing University, Chongqing, China
| | - Wensheng Hou
- Key Laboratory of Biorheological Science and Technology of Ministry of Education, Bioengineering College of Chongqing University, Chongqing, China
- Collaborative Innovation Center for Brain Science, Chongqing University, Chongqing, China
- Chongqing Medical Electronics Engineering Technology Research Center, Chongqing University, Chongqing, China
| |
Collapse
|
25
|
Paraouty N, Stasiak A, Lorenzi C, Varnet L, Winter IM. Dual Coding of Frequency Modulation in the Ventral Cochlear Nucleus. J Neurosci 2018; 38:4123-4137. [PMID: 29599389 PMCID: PMC6596033 DOI: 10.1523/jneurosci.2107-17.2018] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2017] [Revised: 03/18/2018] [Accepted: 03/22/2018] [Indexed: 11/21/2022] Open
Abstract
Frequency modulation (FM) is a common acoustic feature of natural sounds and is known to play a role in robust sound source recognition. Auditory neurons show precise stimulus-synchronized discharge patterns that may be used for the representation of low-rate FM. However, it remains unclear whether this representation is based on synchronization to slow temporal envelope (ENV) cues resulting from cochlear filtering or phase locking to faster temporal fine structure (TFS) cues. To investigate the plausibility of those encoding schemes, single units of the ventral cochlear nucleus of guinea pigs of either sex were recorded in response to sine FM tones centered at the unit's best frequency (BF). The results show that, in contrast to high-BF units, for modulation depths within the receptive field, low-BF units (<4 kHz) demonstrate good phase locking to TFS. For modulation depths extending beyond the receptive field, the discharge patterns follow the ENV and fluctuate at the modulation rate. The receptive field proved to be a good predictor of the ENV responses for most primary-like and chopper units. The current in vivo data also reveal a high level of diversity in responses across unit types. TFS cues are mainly conveyed by low-frequency and primary-like units and ENV cues by chopper and onset units. The diversity of responses exhibited by cochlear nucleus neurons provides a neural basis for a dual-coding scheme of FM in the brainstem based on both ENV and TFS cues.SIGNIFICANCE STATEMENT Natural sounds, including speech, convey informative temporal modulations in frequency. Understanding how the auditory system represents those frequency modulations (FM) has important implications as robust sound source recognition depends crucially on the reception of low-rate FM cues. Here, we recorded 115 single-unit responses from the ventral cochlear nucleus in response to FM and provide the first physiological evidence of a dual-coding mechanism of FM via synchronization to temporal envelope cues and phase locking to temporal fine structure cues. We also demonstrate a diversity of neural responses with different coding specializations. These results support the dual-coding scheme proposed by psychophysicists to account for FM sensitivity in humans and provide new insights on how this might be implemented in the early stages of the auditory pathway.
Collapse
Affiliation(s)
- Nihaad Paraouty
- Centre for the Neural Basis of Hearing, The Physiological Laboratory, Department of Physiology, Development and Neuroscience, University of Cambridge, United Kingdom and
- Laboratoire des Systèmes Perceptifs CNRS UMR 8248, École Normale Supérieure, Paris Sciences et Lettres Research University, Paris, France
| | - Arkadiusz Stasiak
- Centre for the Neural Basis of Hearing, The Physiological Laboratory, Department of Physiology, Development and Neuroscience, University of Cambridge, United Kingdom and
| | - Christian Lorenzi
- Laboratoire des Systèmes Perceptifs CNRS UMR 8248, École Normale Supérieure, Paris Sciences et Lettres Research University, Paris, France
| | - Léo Varnet
- Laboratoire des Systèmes Perceptifs CNRS UMR 8248, École Normale Supérieure, Paris Sciences et Lettres Research University, Paris, France
| | - Ian M Winter
- Centre for the Neural Basis of Hearing, The Physiological Laboratory, Department of Physiology, Development and Neuroscience, University of Cambridge, United Kingdom and
| |
Collapse
|
26
|
Cluster-based analysis improves predictive validity of spike-triggered receptive field estimates. PLoS One 2017; 12:e0183914. [PMID: 28877194 PMCID: PMC5587334 DOI: 10.1371/journal.pone.0183914] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2016] [Accepted: 08/14/2017] [Indexed: 11/19/2022] Open
Abstract
Spectrotemporal receptive field (STRF) characterization is a central goal of auditory physiology. STRFs are often approximated by the spike-triggered average (STA), which reflects the average stimulus preceding a spike. In many cases, the raw STA is subjected to a threshold defined by gain values expected by chance. However, such correction methods have not been universally adopted, and the consequences of specific gain-thresholding approaches have not been investigated systematically. Here, we evaluate two classes of statistical correction techniques, using the resulting STRF estimates to predict responses to a novel validation stimulus. The first, more traditional technique eliminated STRF pixels (time-frequency bins) with gain values expected by chance. This correction method yielded significant increases in prediction accuracy, including when the threshold setting was optimized for each unit. The second technique was a two-step thresholding procedure wherein clusters of contiguous pixels surviving an initial gain threshold were then subjected to a cluster mass threshold based on summed pixel values. This approach significantly improved upon even the best gain-thresholding techniques. Additional analyses suggested that allowing threshold settings to vary independently for excitatory and inhibitory subfields of the STRF resulted in only marginal additional gains, at best. In summary, augmenting reverse correlation techniques with principled statistical correction choices increased prediction accuracy by over 80% for multi-unit STRFs and by over 40% for single-unit STRFs, furthering the interpretational relevance of the recovered spectrotemporal filters for auditory systems analysis.
Collapse
|
27
|
Elmer S, Hausheer M, Albrecht J, Kühnis J. Human Brainstem Exhibits higher Sensitivity and Specificity than Auditory-Related Cortex to Short-Term Phonetic Discrimination Learning. Sci Rep 2017; 7:7455. [PMID: 28785043 PMCID: PMC5547112 DOI: 10.1038/s41598-017-07426-y] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2017] [Accepted: 06/28/2017] [Indexed: 01/09/2023] Open
Abstract
Phonetic discrimination learning is an active perceptual process that operates under the influence of cognitive control mechanisms by increasing the sensitivity of the auditory system to the trained stimulus attributes. It is assumed that the auditory cortex and the brainstem interact in order to refine how sounds are transcribed into neural codes. Here, we evaluated whether these two computational entities are prone to short-term functional changes, whether there is a chronological difference in malleability, and whether short-term training suffices to alter reciprocal interactions. We performed repeated cortical (i.e., mismatch negativity responses, MMN) and subcortical (i.e., frequency-following response, FFR) EEG measurements in two groups of participants who underwent one hour of phonetic discrimination training or were passively exposed to the same stimulus material. The training group showed a distinctive brainstem energy reduction in the trained frequency-range (i.e., first formant), whereas the passive group did not show any response modulation. Notably, brainstem signal change correlated with the behavioral improvement during training, this result indicating a close relationship between behavior and underlying brainstem physiology. Since we did not reveal group differences in MMN responses, results point to specific short-term brainstem changes that precede functional alterations in the auditory cortex.
Collapse
Affiliation(s)
- Stefan Elmer
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Zurich, Switzerland.
| | - Marcela Hausheer
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Zurich, Switzerland
| | - Joëlle Albrecht
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Zurich, Switzerland
| | - Jürg Kühnis
- Auditory Research Group Zurich (ARGZ), Division Neuropsychology, Institute of Psychology, University of Zurich, Zurich, Switzerland
| |
Collapse
|
28
|
Roberts PD, Portfors CV. Responses to Social Vocalizations in the Dorsal Cochlear Nucleus of Mice. Front Syst Neurosci 2015; 9:172. [PMID: 26733824 PMCID: PMC4680083 DOI: 10.3389/fnsys.2015.00172] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2015] [Accepted: 11/26/2015] [Indexed: 11/18/2022] Open
Abstract
Identifying sounds is critical for an animal to make appropriate behavioral responses to environmental stimuli, including vocalizations from conspecifics. Identification of vocalizations may be supported by neuronal selectivity in the auditory pathway. The first place in the ascending auditory pathway where neuronal selectivity to vocalizations has been found is in the inferior colliculus (IC), but very few brainstem nuclei have been evaluated. Here, we tested whether selectivity to vocalizations is present in the dorsal cochlear nucleus (DCN). We recorded extracellular neural responses in the DCN of mice and found that fusiform cells responded in a heterogeneous and selective manner to mouse ultrasonic vocalizations. Most fusiform cells responded to vocalizations that contained spectral energy at much higher frequencies than the characteristic frequencies of the cells. To understand this mismatch of stimulus properties and frequency tuning of the cells, we developed a dynamic, nonlinear model of the cochlea that simulates cochlear distortion products on the basilar membrane. We preprocessed the vocalization stimuli through this model and compared responses to these distorted vocalizations with responses to the original vocalizations. We found that fusiform cells in the DCN respond in a heterogeneous manner to vocalizations, and that these neurons can use distortion products as a mechanism for encoding ultrasonic vocalizations. In addition, the selective neuronal responses were dependent on the presence of inhibitory sidebands that modulated the response depending on the temporal structure of the distortion product. These findings suggest that important processing of complex sounds occurs at a very early stage of central auditory processing and is not strictly a function of the cortex.
Collapse
Affiliation(s)
- Patrick D Roberts
- School of Biological Sciences and Integrative Physiology and Neuroscience, Washington State University Vancouver, WA, USA
| | - Christine V Portfors
- School of Biological Sciences and Integrative Physiology and Neuroscience, Washington State University Vancouver, WA, USA
| |
Collapse
|
29
|
Papesh MA, Hurley LM. Modulation of auditory brainstem responses by serotonin and specific serotonin receptors. Hear Res 2015; 332:121-136. [PMID: 26688176 DOI: 10.1016/j.heares.2015.11.014] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/23/2015] [Revised: 10/28/2015] [Accepted: 11/23/2015] [Indexed: 11/16/2022]
Abstract
The neuromodulator serotonin is found throughout the auditory system from the cochlea to the cortex. Although effects of serotonin have been reported at the level of single neurons in many brainstem nuclei, how these effects correspond to more integrated measures of auditory processing has not been well-explored. In the present study, we aimed to characterize the effects of serotonin on far-field auditory brainstem responses (ABR) across a wide range of stimulus frequencies and intensities. Using a mouse model, we investigated the consequences of systemic serotonin depletion, as well as the selective stimulation and suppression of the 5-HT1 and 5-HT2 receptors, on ABR latency and amplitude. Stimuli included tone pips spanning four octaves presented over a forty dB range. Depletion of serotonin reduced the ABR latencies in Wave II and later waves, suggesting that serotonergic effects occur as early as the cochlear nucleus. Further, agonists and antagonists of specific serotonergic receptors had different profiles of effects on ABR latencies and amplitudes across waves and frequencies, suggestive of distinct effects of these agents on auditory processing. Finally, most serotonergic effects were more pronounced at lower ABR frequencies, suggesting larger or more directional modulation of low-frequency processing. This is the first study to describe the effects of serotonin on ABR responses across a wide range of stimulus frequencies and amplitudes, and it presents an important step in understanding how serotonergic modulation of auditory brainstem processing may contribute to modulation of auditory perception.
Collapse
Affiliation(s)
- Melissa A Papesh
- Indiana University, Department of Speech and Hearing Sciences, 200 South Jordan Avenue, Bloomington, IN 47405, USA.
| | - Laura M Hurley
- Indiana University, Department of Biology, Center for the Integrative Study of Animal Behavior, 1001 E. Third Street, Bloomington, IN 47405, USA
| |
Collapse
|
30
|
Pannese A, Grandjean D, Frühholz S. Subcortical processing in auditory communication. Hear Res 2015; 328:67-77. [DOI: 10.1016/j.heares.2015.07.003] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/03/2015] [Revised: 06/23/2015] [Accepted: 07/01/2015] [Indexed: 12/21/2022]
|
31
|
Rabang CF, Lin J, Wu GK. Balance or imbalance: inhibitory circuits for direction selectivity in the auditory system. Cell Mol Life Sci 2015; 72:1893-906. [PMID: 25638210 PMCID: PMC11113761 DOI: 10.1007/s00018-015-1841-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2014] [Revised: 01/13/2015] [Accepted: 01/15/2015] [Indexed: 10/24/2022]
Abstract
The auditory system detects and processes dynamic sound information transmitted in the environment. Other than the basic acoustic parameters, such as frequency, amplitude and phase, the time-varying changes of these parameters must also be encoded in our brain. Frequency-modulated (FM) sound is socially and environmentally significant, and the direction of FM sweeps is essential for animal communication and human speech. Many auditory neurons selectively respond to the directional change of such FM signals. In the past half century, our knowledge of auditory representation and processing has been updated frequently, due to technological advancement. Recently, in vivo whole-cell voltage clamp recordings have been applied to different brain regions in sensory systems. These recordings illustrate the synaptic mechanisms underlying basic sensory information processing and provide profound insights toward our understanding of neural circuits for complex signal analysis. In this review, we summarize the major findings of direction selectivity at several key auditory regions and emphasize on the recent discoveries on the synaptic mechanisms for direction selectivity in the auditory system. We conclude this review by describing promising technical developments in dissecting neural circuits and future directions in the study of complex sound analysis.
Collapse
Affiliation(s)
- Cal F. Rabang
- Department of Psychology, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
- George Washington Institute for Neuroscience, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
| | - Jeff Lin
- Department of Psychology, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
- George Washington Institute for Neuroscience, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
| | - Guangying K. Wu
- Department of Psychology, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
- George Washington Institute for Neuroscience, The George Washington University, 2300 Eye St NW, Washington, DC 20037 USA
| |
Collapse
|
32
|
Lindeberg T, Friberg A. Idealized computational models for auditory receptive fields. PLoS One 2015; 10:e0119032. [PMID: 25822973 PMCID: PMC4379182 DOI: 10.1371/journal.pone.0119032] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2014] [Accepted: 01/24/2015] [Indexed: 11/19/2022] Open
Abstract
We present a theory by which idealized models of auditory receptive fields can be derived in a principled axiomatic manner, from a set of structural properties to (i) enable invariance of receptive field responses under natural sound transformations and (ii) ensure internal consistency between spectro-temporal receptive fields at different temporal and spectral scales. For defining a time-frequency transformation of a purely temporal sound signal, it is shown that the framework allows for a new way of deriving the Gabor and Gammatone filters as well as a novel family of generalized Gammatone filters, with additional degrees of freedom to obtain different trade-offs between the spectral selectivity and the temporal delay of time-causal temporal window functions. When applied to the definition of a second-layer of receptive fields from a spectrogram, it is shown that the framework leads to two canonical families of spectro-temporal receptive fields, in terms of spectro-temporal derivatives of either spectro-temporal Gaussian kernels for non-causal time or a cascade of time-causal first-order integrators over the temporal domain and a Gaussian filter over the logspectral domain. For each filter family, the spectro-temporal receptive fields can be either separable over the time-frequency domain or be adapted to local glissando transformations that represent variations in logarithmic frequencies over time. Within each domain of either non-causal or time-causal time, these receptive field families are derived by uniqueness from the assumptions. It is demonstrated how the presented framework allows for computation of basic auditory features for audio processing and that it leads to predictions about auditory receptive fields with good qualitative similarity to biological receptive fields measured in the inferior colliculus (ICC) and primary auditory cortex (A1) of mammals.
Collapse
Affiliation(s)
- Tony Lindeberg
- Department of Computational Biology, School of Computer Science and Communication, KTH Royal Institute of Technology, Stockholm, Sweden
| | - Anders Friberg
- Department of Speech, Music and Hearing, School of Computer Science and Communication, KTH Royal Institute of Technology, Stockholm, Sweden
| |
Collapse
|
33
|
Spectrotemporal response properties of core auditory cortex neurons in awake monkey. PLoS One 2015; 10:e0116118. [PMID: 25680187 PMCID: PMC4332665 DOI: 10.1371/journal.pone.0116118] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2014] [Accepted: 12/03/2014] [Indexed: 11/19/2022] Open
Abstract
So far, most studies of core auditory cortex (AC) have characterized the spectral and temporal tuning properties of cells in non-awake, anesthetized preparations. As experiments in awake animals are scarce, we here used dynamic spectral-temporal broadband ripples to study the properties of the spectrotemporal receptive fields (STRFs) of AC cells in awake monkeys. We show that AC neurons were typically most sensitive to low ripple densities (spectral) and low velocities (temporal), and that most cells were not selective for a particular spectrotemporal sweep direction. A substantial proportion of neurons preferred amplitude-modulated sounds (at zero ripple density) to dynamic ripples (at non-zero densities). The vast majority (>93%) of modulation transfer functions were separable with respect to spectral and temporal modulations, indicating that time and spectrum are independently processed in AC neurons. We also analyzed the linear predictability of AC responses to natural vocalizations on the basis of the STRF. We discuss our findings in the light of results obtained from the monkey midbrain inferior colliculus by comparing the spectrotemporal tuning properties and linear predictability of these two important auditory stages.
Collapse
|
34
|
|
35
|
Meyer AF, Diepenbrock JP, Happel MFK, Ohl FW, Anemüller J. Discriminative learning of receptive fields from responses to non-Gaussian stimulus ensembles. PLoS One 2014; 9:e93062. [PMID: 24699631 PMCID: PMC3974709 DOI: 10.1371/journal.pone.0093062] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2013] [Accepted: 02/28/2014] [Indexed: 11/19/2022] Open
Abstract
Analysis of sensory neurons' processing characteristics requires simultaneous measurement of presented stimuli and concurrent spike responses. The functional transformation from high-dimensional stimulus space to the binary space of spike and non-spike responses is commonly described with linear-nonlinear models, whose linear filter component describes the neuron's receptive field. From a machine learning perspective, this corresponds to the binary classification problem of discriminating spike-eliciting from non-spike-eliciting stimulus examples. The classification-based receptive field (CbRF) estimation method proposed here adapts a linear large-margin classifier to optimally predict experimental stimulus-response data and subsequently interprets learned classifier weights as the neuron's receptive field filter. Computational learning theory provides a theoretical framework for learning from data and guarantees optimality in the sense that the risk of erroneously assigning a spike-eliciting stimulus example to the non-spike class (and vice versa) is minimized. Efficacy of the CbRF method is validated with simulations and for auditory spectro-temporal receptive field (STRF) estimation from experimental recordings in the auditory midbrain of Mongolian gerbils. Acoustic stimulation is performed with frequency-modulated tone complexes that mimic properties of natural stimuli, specifically non-Gaussian amplitude distribution and higher-order correlations. Results demonstrate that the proposed approach successfully identifies correct underlying STRFs, even in cases where second-order methods based on the spike-triggered average (STA) do not. Applied to small data samples, the method is shown to converge on smaller amounts of experimental recordings and with lower estimation variance than the generalized linear model and recent information theoretic methods. Thus, CbRF estimation may prove useful for investigation of neuronal processes in response to natural stimuli and in settings where rapid adaptation is induced by experimental design.
Collapse
Affiliation(s)
- Arne F. Meyer
- Department of Medical Physics and Acoustics and Cluster of Excellence ''Hearing4all'', University of Oldenburg, Oldenburg, Germany
- * E-mail:
| | - Jan-Philipp Diepenbrock
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
| | - Max F. K. Happel
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
- Department of Neuroprosthetics, Institute of Biology, Otto-von-Guericke University, Magdeburg, Germany
| | - Frank W. Ohl
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
- Department of Neuroprosthetics, Institute of Biology, Otto-von-Guericke University, Magdeburg, Germany
| | - Jörn Anemüller
- Department of Medical Physics and Acoustics and Cluster of Excellence ''Hearing4all'', University of Oldenburg, Oldenburg, Germany
| |
Collapse
|
36
|
Binaural gain modulation of spectrotemporal tuning in the interaural level difference-coding pathway. J Neurosci 2013; 33:11089-99. [PMID: 23825414 DOI: 10.1523/jneurosci.4941-12.2013] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
In the brainstem, the auditory system diverges into two pathways that process different sound localization cues, interaural time differences (ITDs) and level differences (ILDs). We investigated the site where ILD is detected in the auditory system of barn owls, the posterior part of the lateral lemniscus (LLDp). This structure is equivalent to the lateral superior olive in mammals. The LLDp is unique in that it is the first place of binaural convergence in the brainstem where monaural excitatory and inhibitory inputs converge. Using binaurally uncorrelated noise and a generalized linear model, we were able to estimate the spectrotemporal tuning of excitatory and inhibitory inputs to these cells. We show that the response of LLDp neurons is highly locked to the stimulus envelope. Our data demonstrate that spectrotemporally tuned, temporally delayed inhibition enhances the reliability of envelope locking by modulating the gain of LLDp neurons' responses. The dependence of gain modulation on ILD shown here constitutes a means for space-dependent coding of stimulus identity by the initial stages of the auditory pathway.
Collapse
|
37
|
Honma Y, Tsukano H, Horie M, Ohshima S, Tohmi M, Kubota Y, Takahashi K, Hishida R, Takahashi S, Shibuki K. Auditory cortical areas activated by slow frequency-modulated sounds in mice. PLoS One 2013; 8:e68113. [PMID: 23874516 PMCID: PMC3714279 DOI: 10.1371/journal.pone.0068113] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2013] [Accepted: 05/24/2013] [Indexed: 11/18/2022] Open
Abstract
Species-specific vocalizations in mice have frequency-modulated (FM) components slower than the lower limit of FM direction selectivity in the core region of the mouse auditory cortex. To identify cortical areas selective to slow frequency modulation, we investigated tonal responses in the mouse auditory cortex using transcranial flavoprotein fluorescence imaging. For differentiating responses to frequency modulation from those to stimuli at constant frequencies, we focused on transient fluorescence changes after direction reversal of temporally repeated and superimposed FM sweeps. We found that the ultrasonic field (UF) in the belt cortical region selectively responded to the direction reversal. The dorsoposterior field (DP) also responded weakly to the reversal. Regarding the responses in UF, no apparent tonotopic map was found, and the right UF responses were significantly larger in amplitude than the left UF responses. The half-max latency in responses to FM sweeps was shorter in UF compared with that in the primary auditory cortex (A1) or anterior auditory field (AAF). Tracer injection experiments in the functionally identified UF and DP confirmed that these two areas receive afferent inputs from the dorsal part of the medial geniculate nucleus (MG). Calcium imaging of UF neurons stained with fura-2 were performed using a two-photon microscope, and the presence of UF neurons that were selective to both direction and direction reversal of slow frequency modulation was demonstrated. These results strongly suggest a role for UF, and possibly DP, as cortical areas specialized for processing slow frequency modulation in mice.
Collapse
Affiliation(s)
- Yuusuke Honma
- Department of Neurophysiology, Brain Research Institute, Niigata University, Asahi-machi, Niigata, Japan
- Department of Otolaryngology, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Hiroaki Tsukano
- Department of Neurophysiology, Brain Research Institute, Niigata University, Asahi-machi, Niigata, Japan
| | - Masao Horie
- Department of Anatomy, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Shinsuke Ohshima
- Department of Otolaryngology, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Manavu Tohmi
- Department of Neurophysiology, Brain Research Institute, Niigata University, Asahi-machi, Niigata, Japan
| | - Yamato Kubota
- Department of Otolaryngology, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Kuniyuki Takahashi
- Department of Otolaryngology, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Ryuichi Hishida
- Department of Neurophysiology, Brain Research Institute, Niigata University, Asahi-machi, Niigata, Japan
| | - Sugata Takahashi
- Department of Otolaryngology, Faculty of Medicine, Niigata University, Asahi-machi, Niigata, Japan
| | - Katsuei Shibuki
- Department of Neurophysiology, Brain Research Institute, Niigata University, Asahi-machi, Niigata, Japan
- * E-mail:
| |
Collapse
|
38
|
Conserved mechanisms of vocalization coding in mammalian and songbird auditory midbrain. Hear Res 2013; 305:45-56. [PMID: 23726970 DOI: 10.1016/j.heares.2013.05.005] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/17/2012] [Revised: 03/23/2013] [Accepted: 05/11/2013] [Indexed: 11/23/2022]
Abstract
The ubiquity of social vocalizations among animals provides the opportunity to identify conserved mechanisms of auditory processing that subserve communication. Identifying auditory coding properties that are shared across vocal communicators will provide insight into how human auditory processing leads to speech perception. Here, we compare auditory response properties and neural coding of social vocalizations in auditory midbrain neurons of mammalian and avian vocal communicators. The auditory midbrain is a nexus of auditory processing because it receives and integrates information from multiple parallel pathways and provides the ascending auditory input to the thalamus. The auditory midbrain is also the first region in the ascending auditory system where neurons show complex tuning properties that are correlated with the acoustics of social vocalizations. Single unit studies in mice, bats and zebra finches reveal shared principles of auditory coding including tonotopy, excitatory and inhibitory interactions that shape responses to vocal signals, nonlinear response properties that are important for auditory coding of social vocalizations and modulation tuning. Additionally, single neuron responses in the mouse and songbird midbrain are reliable, selective for specific syllables, and rely on spike timing for neural discrimination of distinct vocalizations. We propose that future research on auditory coding of vocalizations in mouse and songbird midbrain neurons adopt similar experimental and analytical approaches so that conserved principles of vocalization coding may be distinguished from those that are specialized for each species. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives".
Collapse
|
39
|
Pollak GD. The dominant role of inhibition in creating response selectivities for communication calls in the brainstem auditory system. Hear Res 2013; 305:86-101. [PMID: 23545427 DOI: 10.1016/j.heares.2013.03.001] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/12/2012] [Revised: 02/20/2013] [Accepted: 03/06/2013] [Indexed: 10/27/2022]
Abstract
This review is concerned with how communication calls are processed and represented by populations of neurons in both the inferior colliculus (IC), the auditory midbrain nucleus, and the dorsal nucleus of the lateral lemniscus (DNLL), the nucleus just caudal to the IC. The review has five sections where focus in each section is on inhibition and its role in shaping response selectivity for communication calls. In the first section, the lack of response selectivity for calls in DNLL neurons is presented and discusses why inhibition plays virtually no role in shaping selectivity. In the second section, the lack of selectivity in the DNLL is contrasted with the high degree of response selectivity in the IC. The third section then reviews how inhibition in the IC shapes response selectivities for calls, and how those selectivities can create a population response with a distinctive response profile to a particular call, which differs from the population profile evoked by any other call. The fourth section is concerned with the specifics of inhibition in the IC, and how the interaction of excitation and inhibition creates directional selectivities for frequency modulations, one of the principal acoustic features of communication signals. The two major hypotheses for directional selectivity are presented. One is the timing hypothesis, which holds that the precise timing of excitation relative to inhibition is the feature that shapes directionality. The other hypothesis is that the relative magnitudes of excitation and inhibition are the dominant features that shape directionality, where timing is relatively unimportant. The final section then turns to the role of serotonin, a neuromodulator that can markedly change responses to calls in the IC. Serotonin provides a linkage between behavioral states and processing. This linkage is discussed in the final section together with the hypothesis that serotonin acts to enhances the contrast in the population responses to various calls over and above the distinctive population responses that were created by inhibition. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives".
Collapse
Affiliation(s)
- George D Pollak
- Section of Neurobiology and Center for Perceptual Systems, 337 Patterson Laboratory Building, The University of Texas at Austin, Austin, TX 78712, USA.
| |
Collapse
|
40
|
Felix RA, Vonderschen K, Berrebi AS, Magnusson AK. Development of on-off spiking in superior paraolivary nucleus neurons of the mouse. J Neurophysiol 2013; 109:2691-704. [PMID: 23515791 DOI: 10.1152/jn.01041.2012] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
The superior paraolivary nucleus (SPON) is a prominent cell group in the auditory brain stem that has been increasingly implicated in representing temporal sound structure. Although SPON neurons selectively respond to acoustic signals important for sound periodicity, the underlying physiological specializations enabling these responses are poorly understood. We used in vitro and in vivo recordings to investigate how SPON neurons develop intrinsic cellular properties that make them well suited for encoding temporal sound features. In addition to their hallmark rebound spiking at the stimulus offset, SPON neurons were characterized by spiking patterns termed onset, adapting, and burst in response to depolarizing stimuli in vitro. Cells with burst spiking had some morphological differences compared with other SPON neurons and were localized to the dorsolateral region of the nucleus. Both membrane and spiking properties underwent strong developmental regulation, becoming more temporally precise with age for both onset and offset spiking. Single-unit recordings obtained in young mice demonstrated that SPON neurons respond with temporally precise onset spiking upon tone stimulation in vivo, in addition to the typical offset spiking. Taken together, the results of the present study demonstrate that SPON neurons develop sharp on-off spiking, which may confer sensitivity to sound amplitude modulations or abrupt sound transients. These findings are consistent with the proposed involvement of the SPON in the processing of temporal sound structure, relevant for encoding communication cues.
Collapse
Affiliation(s)
- Richard A Felix
- Center for Hearing and Communication Research, Karolinska Institutet and Department of Clinical Science, Intervention and Technology, Karolinska University Hospital, Stockholm, Sweden
| | | | | | | |
Collapse
|
41
|
Geis HRAP, Borst JGG. Intracellular responses to frequency modulated tones in the dorsal cortex of the mouse inferior colliculus. Front Neural Circuits 2013; 7:7. [PMID: 23386812 PMCID: PMC3560375 DOI: 10.3389/fncir.2013.00007] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2012] [Accepted: 01/13/2013] [Indexed: 11/13/2022] Open
Abstract
Frequency modulations occur in many natural sounds, including vocalizations. The neuronal response to frequency modulated (FM) stimuli has been studied extensively in different brain areas, with an emphasis on the auditory cortex and the central nucleus of the inferior colliculus. Here, we measured the responses to FM sweeps in whole-cell recordings from neurons in the dorsal cortex of the mouse inferior colliculus. Both up- and downward logarithmic FM sweeps were presented at two different speeds to both the ipsi- and the contralateral ear. Based on the number of action potentials that were fired, between 10 and 24% of cells were selective for rate or direction of the FM sweeps. A somewhat lower percentage of cells, 6–21%, showed selectivity based on EPSP size. To study the mechanisms underlying the generation of FM selectivity, we compared FM responses with responses to simple tones in the same cells. We found that if pairs of neurons responded in a similar way to simple tones, they generally also responded in a similar way to FM sweeps. Further evidence that FM selectivity can be generated within the dorsal cortex was obtained by reconstructing FM sweeps from the response to simple tones using three different models. In about half of the direction selective neurons the selectivity was generated by spectrally asymmetric synaptic inhibition. In addition, evidence for direction selectivity based on the timing of excitatory responses was also obtained in some cells. No clear evidence for the local generation of rate selectivity was obtained. We conclude that FM direction selectivity can be generated within the dorsal cortex of the mouse inferior colliculus by multiple mechanisms.
Collapse
Affiliation(s)
- H-Rüdiger A P Geis
- Department of Neuroscience, Erasmus MC, University Medical Center Rotterdam Rotterdam, Netherlands
| | | |
Collapse
|
42
|
Laudanski J, Edeline JM, Huetz C. Differences between spectro-temporal receptive fields derived from artificial and natural stimuli in the auditory cortex. PLoS One 2012; 7:e50539. [PMID: 23209771 PMCID: PMC3507792 DOI: 10.1371/journal.pone.0050539] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2012] [Accepted: 10/25/2012] [Indexed: 11/25/2022] Open
Abstract
Spectro-temporal properties of auditory cortex neurons have been extensively studied with artificial sounds but it is still unclear whether they help in understanding neuronal responses to communication sounds. Here, we directly compared spectro-temporal receptive fields (STRFs) obtained from the same neurons using both artificial stimuli (dynamic moving ripples, DMRs) and natural stimuli (conspecific vocalizations) that were matched in terms of spectral content, average power and modulation spectrum. On a population of auditory cortex neurons exhibiting reliable tuning curves when tested with pure tones, significant STRFs were obtained for 62% of the cells with vocalizations and 68% with DMR. However, for many cells with significant vocalization-derived STRFs (STRFvoc) and DMR-derived STRFs (STRFdmr), the BF, latency, bandwidth and global STRFs shape differed more than what would be predicted by spiking responses simulated by a linear model based on a non-homogenous Poisson process. Moreover STRFvoc predicted neural responses to vocalizations more accurately than STRFdmr predicted neural response to DMRs, despite similar spike-timing reliability for both sets of stimuli. Cortical bursts, which potentially introduce nonlinearities in evoked responses, did not explain the differences between STRFvoc and STRFdmr. Altogether, these results suggest that the nonlinearity of auditory cortical responses makes it difficult to predict responses to communication sounds from STRFs computed from artificial stimuli.
Collapse
Affiliation(s)
- Jonathan Laudanski
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| | - Jean-Marc Edeline
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
- * E-mail:
| | - Chloé Huetz
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| |
Collapse
|
43
|
Thompson JV, Jeanne JM, Gentner TQ. Local inhibition modulates learning-dependent song encoding in the songbird auditory cortex. J Neurophysiol 2012; 109:721-33. [PMID: 23155175 DOI: 10.1152/jn.00262.2012] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Changes in inhibition during development are well documented, but the role of inhibition in adult learning-related plasticity is not understood. In songbirds, vocal recognition learning alters the neural representation of songs across the auditory forebrain, including the caudomedial nidopallium (NCM), a region analogous to mammalian secondary auditory cortices. Here, we block local inhibition with the iontophoretic application of gabazine, while simultaneously measuring song-evoked spiking activity in NCM of European starlings trained to recognize sets of conspecific songs. We find that local inhibition differentially suppresses the responses to learned and unfamiliar songs and enhances spike-rate differences between learned categories of songs. These learning-dependent response patterns emerge, in part, through inhibitory modulation of selectivity for song components and the masking of responses to specific acoustic features without altering spectrotemporal tuning. The results describe a novel form of inhibitory modulation of the encoding of learned categories and demonstrate that inhibition plays a central role in shaping the responses of neurons to learned, natural signals.
Collapse
Affiliation(s)
- Jason V Thompson
- Neurosciences Graduate Program, University of California, San Diego, La Jolla, CA, USA
| | | | | |
Collapse
|
44
|
Gittelman JX, Wang L, Colburn HS, Pollak GD. Inhibition shapes response selectivity in the inferior colliculus by gain modulation. Front Neural Circuits 2012; 6:67. [PMID: 23024629 PMCID: PMC3444759 DOI: 10.3389/fncir.2012.00067] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2012] [Accepted: 08/31/2012] [Indexed: 12/20/2022] Open
Abstract
Pharmacological block of inhibition is often used to determine if inhibition contributes to spike selectivity, in which a preferred stimulus evokes more spikes than a null stimulus. When inhibitory block reduces spike selectivity, a common interpretation is that differences between the preferred- and null-evoked inhibitions created the selectivity from less-selective excitatory inputs. In models based on empirical properties of cells from the inferior colliculus (IC) of awake bats, we show that inhibitory differences are not required. Instead, inhibition can enhance spike selectivity by changing the gain, the ratio of output spikes to input current. Within the model, we made preferred stimuli that evoked more spikes than null stimuli using five distinct synaptic mechanisms. In two cases, synaptic selectivity (the differences between the preferred and null inputs) was entirely excitatory, and in two it was entirely inhibitory. In each case, blocking inhibition eliminated spike selectivity. Thus, observing spike rates following inhibitory block did not distinguish among the cases where synaptic selectivity was entirely excitatory or inhibitory. We then did the same modeling experiment using empirical synaptic conductances derived from responses to preferred and null sounds. In most cases, inhibition in the model enhanced spike selectivity mainly by gain modulation and firing rate reduction. Sometimes, inhibition reduced the null gain to zero, eliminating null-evoked spikes. In some cases, inhibition increased the preferred gain more than the null gain, enhancing the difference between the preferred- and null-evoked spikes. Finally, inhibition kept firing rates low. When selectivity is quantified by the selectivity index (SI, the ratio of the difference to the sum of the spikes evoked by the preferred and null stimuli), inhibitory block reduced the SI by increasing overall firing rates. These results are consistent with inhibition shaping spike selectivity by gain control.
Collapse
Affiliation(s)
- Joshua X Gittelman
- Section of Neurobiology, Institute for Neuroscience, Center for Perceptual Systems, The University of Texas Austin, TX, USA
| | | | | | | |
Collapse
|
45
|
Hurley LM, Sullivan MR. From behavioral context to receptors: serotonergic modulatory pathways in the IC. Front Neural Circuits 2012; 6:58. [PMID: 22973195 PMCID: PMC3434355 DOI: 10.3389/fncir.2012.00058] [Citation(s) in RCA: 43] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2012] [Accepted: 08/10/2012] [Indexed: 12/18/2022] Open
Abstract
In addition to ascending, descending, and lateral auditory projections, inputs extrinsic to the auditory system also influence neural processing in the inferior colliculus (IC). These types of inputs often have an important role in signaling salient factors such as behavioral context or internal state. One route for such extrinsic information is through centralized neuromodulatory networks like the serotonergic system. Serotonergic inputs to the IC originate from centralized raphe nuclei, release serotonin in the IC, and activate serotonin receptors expressed by auditory neurons. Different types of serotonin receptors act as parallel pathways regulating specific features of circuitry within the IC. This results from variation in subcellular localizations and effector pathways of different receptors, which consequently influence auditory responses in distinct ways. Serotonin receptors may regulate GABAergic inhibition, influence response gain, alter spike timing, or have effects that are dependent on the level of activity. Serotonin receptor types additionally interact in nonadditive ways to produce distinct combinatorial effects. This array of effects of serotonin is likely to depend on behavioral context, since the levels of serotonin in the IC transiently increase during behavioral events including stressful situations and social interaction. These studies support a broad model of serotonin receptors as a link between behavioral context and reconfiguration of circuitry in the IC, and the resulting possibility that plasticity at the level of specific receptor types could alter the relationship between context and circuit function.
Collapse
Affiliation(s)
- Laura M Hurley
- Department of Biology, Center for the Integrative Study of Animal Behavior, Indiana University Bloomington, IN, USA
| | | |
Collapse
|
46
|
Carlson NL, Ming VL, DeWeese MR. Sparse codes for speech predict spectrotemporal receptive fields in the inferior colliculus. PLoS Comput Biol 2012; 8:e1002594. [PMID: 22807665 PMCID: PMC3395612 DOI: 10.1371/journal.pcbi.1002594] [Citation(s) in RCA: 41] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2011] [Accepted: 05/18/2012] [Indexed: 12/03/2022] Open
Abstract
We have developed a sparse mathematical representation of speech that minimizes the number of active model neurons needed to represent typical speech sounds. The model learns several well-known acoustic features of speech such as harmonic stacks, formants, onsets and terminations, but we also find more exotic structures in the spectrogram representation of sound such as localized checkerboard patterns and frequency-modulated excitatory subregions flanked by suppressive sidebands. Moreover, several of these novel features resemble neuronal receptive fields reported in the Inferior Colliculus (IC), as well as auditory thalamus and cortex, and our model neurons exhibit the same tradeoff in spectrotemporal resolution as has been observed in IC. To our knowledge, this is the first demonstration that receptive fields of neurons in the ascending mammalian auditory pathway beyond the auditory nerve can be predicted based on coding principles and the statistical properties of recorded sounds.
Collapse
Affiliation(s)
- Nicole L. Carlson
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
- Department of Physics, University of California, Berkeley, California, United States of America
| | - Vivienne L. Ming
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
| | - Michael Robert DeWeese
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
- Department of Physics, University of California, Berkeley, California, United States of America
- Helen Wills Neuroscience Institute, University of California, Berkeley, California, United States of America
| |
Collapse
|
47
|
Hsieh IH, Fillmore P, Rong F, Hickok G, Saberi K. FM-selective networks in human auditory cortex revealed using fMRI and multivariate pattern classification. J Cogn Neurosci 2012; 24:1896-907. [PMID: 22640390 DOI: 10.1162/jocn_a_00254] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Frequency modulation (FM) is an acoustic feature of nearly all complex sounds. Directional FM sweeps are especially pervasive in speech, music, animal vocalizations, and other natural sounds. Although the existence of FM-selective cells in the auditory cortex of animals has been documented, evidence in humans remains equivocal. Here we used multivariate pattern analysis to identify cortical selectivity for direction of a multitone FM sweep. This method distinguishes one pattern of neural activity from another within the same ROI, even when overall level of activity is similar, allowing for direct identification of FM-specialized networks. Standard contrast analysis showed that despite robust activity in auditory cortex, no clusters of activity were associated with up versus down sweeps. Multivariate pattern analysis classification, however, identified two brain regions as selective for FM direction, the right primary auditory cortex on the supratemporal plane and the left anterior region of the superior temporal gyrus. These findings are the first to directly demonstrate existence of FM direction selectivity in the human auditory cortex.
Collapse
Affiliation(s)
- I-Hui Hsieh
- National Central University, Jhongli City, Taiwan.
| | | | | | | | | |
Collapse
|
48
|
Kuo RI, Wu GK. The generation of direction selectivity in the auditory system. Neuron 2012; 73:1016-27. [PMID: 22405210 DOI: 10.1016/j.neuron.2011.11.035] [Citation(s) in RCA: 51] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/29/2011] [Indexed: 01/10/2023]
Abstract
Both human speech and animal vocal signals contain frequency-modulated (FM) sounds. Although central auditory neurons that selectively respond to the direction of frequency modulation are known, the synaptic mechanisms underlying the generation of direction selectivity (DS) remain elusive. Here we show the emergence of DS neurons in the inferior colliculus by mapping the three major subcortical auditory nuclei. Cell-attached recordings reveal a highly reliable and precise firing of DS neurons to FM sweeps in a preferred direction. By using in vivo whole-cell current-clamp and voltage-clamp recordings, we found that the synaptic inputs to DS neurons are not direction selective, but temporally reversed excitatory and inhibitory synaptic inputs are evoked in response to opposing directions of FM sweeps. The construction of such temporal asymmetry, resulting DS, and its topography can be attributed to the spectral disparity of the excitatory and the inhibitory synaptic tonal receptive fields.
Collapse
Affiliation(s)
- Richard I Kuo
- Broad Fellows Program in Brain Circuitry and Division of Biology, California Institute of Technology, 1200 East California Boulevard, Pasadena, CA 91125, USA
| | | |
Collapse
|
49
|
Atencio CA, Sharpee TO, Schreiner CE. Receptive field dimensionality increases from the auditory midbrain to cortex. J Neurophysiol 2012; 107:2594-603. [PMID: 22323634 DOI: 10.1152/jn.01025.2011] [Citation(s) in RCA: 45] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
In the primary auditory cortex, spectrotemporal receptive fields (STRFs) are composed of multiple independent components that capture the processing of disparate stimulus aspects by any given neuron. The origin of these multidimensional stimulus filters in the central auditory system is unknown. To determine whether multicomponent STRFs emerge prior to the forebrain, we recorded from single neurons in the main obligatory station of the auditory midbrain, the inferior colliculus. By comparing results of different spike-triggered techniques, we found that the neural responses in the inferior colliculus can be accounted for by a single stimulus filter. This was observed for all temporal response patterns, from strongly phasic to tonic. Our results reveal that spectrotemporal stimulus encoding undergoes a fundamental transformation along the auditory neuraxis, with the emergence of multidimensional receptive fields beyond the auditory midbrain.
Collapse
Affiliation(s)
- Craig A Atencio
- The UCSF Center for Integrative Neuroscience, Department of Otolaryngology-Headand Neck Surgery, University of California, San Francisco, CA, USA.
| | | | | |
Collapse
|
50
|
Selectivity for spectral motion as a neural computation for encoding natural communication signals in bat inferior colliculus. J Neurosci 2012; 31:16529-40. [PMID: 22090479 DOI: 10.1523/jneurosci.1306-11.2011] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
This study examines the neural computations performed by neurons in the auditory system to be selective for the direction and velocity of signals sweeping upward or downward in frequency, termed spectral motion. We show that neurons in the auditory midbrain of Mexican free-tailed bats encode multiple spectrotemporal features of natural communication sounds. These features to which each neuron is tuned are nonlinearly combined to produce selectivity for spectral motion cues present in their conspecific calls, such as direction and velocity. We find that the neural computations resulting in selectivity for spectral motion are analogous to models of motion selectivity studied in vision. Our analysis revealed that auditory neurons in the inferior colliculus (IC) are avoiding spectrotemporal modulations that are redundant across different bat communication signals and are specifically tuned for modulations that distinguish each call from another by their frequency-modulated direction and velocity, suggesting that spectral motion is the neural computation through which IC neurons are encoding specific features of conspecific vocalizations.
Collapse
|