1
|
Lyamzin DR, Alamia A, Abdolrahmani M, Aoki R, Benucci A. Regularizing hyperparameters of interacting neural signals in the mouse cortex reflect states of arousal. PLoS Comput Biol 2024; 20:e1012478. [PMID: 39405361 PMCID: PMC11527387 DOI: 10.1371/journal.pcbi.1012478] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2022] [Revised: 10/31/2024] [Accepted: 09/11/2024] [Indexed: 11/02/2024] Open
Abstract
In natural behaviors, multiple neural signals simultaneously drive activation across overlapping brain networks. Due to limitations in the amount of data that can be acquired in common experimental designs, the determination of these interactions is commonly inferred via modeling approaches, which reduce overfitting by finding appropriate regularizing hyperparameters. However, it is unclear whether these hyperparameters can also be related to any aspect of the underlying biological phenomena and help interpret them. We applied a state-of-the-art regularization procedure-automatic locality determination-to interacting neural activations in the mouse posterior cortex associated with movements of the body and eyes. As expected, regularization significantly improved the determination and interpretability of the response interactions. However, regularizing hyperparameters also changed considerably, and seemingly unpredictably, from animal to animal. We found that these variations were not random; rather, they correlated with the variability in visually evoked responses and with the variability in the state of arousal of the animals measured by pupillometry-both pieces of information that were not included in the modeling framework. These observations could be generalized to another commonly used-but potentially less informative-regularization method, ridge regression. Our findings demonstrate that optimal model hyperparameters can be discovery tools that are informative of factors not a priori included in the model's design.
Collapse
Affiliation(s)
| | - Andrea Alamia
- Centre de Recherche Cerveau et Cognition, CNRS, Université de Toulouse, Toulouse, France
| | | | - Ryo Aoki
- RIKEN Center for Brain Science, Wako-shi, Saitama, Japan
| | - Andrea Benucci
- Queen Mary University of London, School of Biological and Behavioural Sciences, London, United Kingdom
| |
Collapse
|
2
|
Strong energy component is more important than spectral selectivity in modeling responses of midbrain auditory neurons to wide-band environmental sounds. Biosystems 2022; 221:104752. [PMID: 36028002 DOI: 10.1016/j.biosystems.2022.104752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2022] [Revised: 07/08/2022] [Accepted: 07/31/2022] [Indexed: 11/23/2022]
Abstract
Modeling central auditory neurons in response to complex sounds not only helps understanding neural processing of speech signals but can also provide insights for biomimetics in neuro-engineering. While modeling responses of midbrain auditory neurons to synthetic tones is rather good, modeling those to environmental sounds is less satisfactory. Environmental sounds typically contain a wide range of frequency components, often with strong and transient energy. These stimulus features have not been examined in the conventional approach of auditory modeling centered on spectral selectivity. To this end, we firstly compared responses to an environmental sound of auditory midbrain neurons across 3 subpopulations of neurons with frequency selectivity in the low, middle and high ranges; secondly, we manipulated the sound energy, both in power and in spectrum, and compared across these subpopulations how their modeled responses were affected. The environmental sound was recorded when a rat was drinking from a feeding bottle (called the 'drinking sound'). The sound spectrum was divided into 20 non-overlapping frequency bands (from 0 to 20 kHz, at 1 kHz width) and presented to an artificial neural model built on a committee machine with parallel spectral inputs to simulate the known tonotopic organization of the auditory system. The model was trained to predict empirical response probability profiles of neurons to the repeated sounds. Results showed that model performance depended more on the strong energy components than on the spectral selectivity. Findings were interpreted to reflect general sensitivity to rapidly changing sound intensities at the auditory midbrain and in the cortex.
Collapse
|
3
|
Schottdorf M, Lee BB. A quantitative description of macaque ganglion cell responses to natural scenes: the interplay of time and space. J Physiol 2021; 599:3169-3193. [PMID: 33913164 DOI: 10.1113/jp281200] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2020] [Accepted: 04/20/2021] [Indexed: 11/08/2022] Open
Abstract
KEY POINTS Responses to natural scenes are the business of the retina. We find primate ganglion cell responses to such scenes consistent with those to simpler stimuli. A biophysical model confirmed this and predicted ganglion cell responses with close to retinal reliability. Primate ganglion cell responses to natural scenes were driven by temporal variations in colour and luminance over the receptive field centre caused by eye movements, and little influenced by interaction of centre and surround with structure in the scene. We discuss implications in the context of efficient coding of the visual environment. Much information in a higher spatiotemporal frequency band is concentrated in the magnocellular pathway. ABSTRACT Responses of visual neurons to natural scenes provide a link between classical descriptions of receptive field structure and visual perception of the natural environment. A natural scene video with a movement pattern resembling that of primate eye movements was used to evoke responses from macaque ganglion cells. Cell responses were well described through known properties of cell receptive fields. Different analyses converge to show that responses primarily derive from the temporal pattern of stimulation derived from eye movements, rather than spatial receptive field structure beyond centre size and position. This was confirmed using a model that predicted ganglion cell responses close to retinal reliability, with only a small contribution of the surround relative to the centre. We also found that the spatiotemporal spectrum of the stimulus is modified in ganglion cell responses, and this can reduce redundancy in the retinal signal. This is more pronounced in the magnocellular pathway, which is much better suited to transmit the detailed structure of natural scenes than the parvocellular pathway. Whitening is less important for chromatic channels. Taken together, this shows how a complex interplay across space, time and spectral content sculpts ganglion cell responses.
Collapse
Affiliation(s)
- Manuel Schottdorf
- Max Planck Institute for Dynamics and Self-Organization, Göttingen, D-37077, Germany.,Max Planck Institute of Experimental Medicine, Göttingen, D-37075, Germany.,Princeton Neuroscience Institute, Princeton, NJ, 08544, USA
| | - Barry B Lee
- Graduate Center for Vision Research, Department of Biological Sciences, SUNY College of Optometry, 33 West 42nd St., New York, NY, 10036, USA.,Department of Neurobiology, Max Planck Institute for Biophysical Chemistry, Göttingen, D-37077, Germany
| |
Collapse
|
4
|
Qureshi F, Yan J. Three dimensional rendering of auditory neuronal responses: A novel illustration of receptive field across frequency, intensity & time domains. J Neurosci Methods 2020; 338:108682. [PMID: 32165230 DOI: 10.1016/j.jneumeth.2020.108682] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2019] [Revised: 03/07/2020] [Accepted: 03/08/2020] [Indexed: 10/24/2022]
Abstract
BACKGROUND Neural coding of sound information is often studied through frequency tuning curve (FTC), spectro-temporal receptive field (STRF), post-stimulus time histogram (PSTH), and other methods such as rate functions. These methods, despite providing a robust characterization of auditory responses in their specific domains, lack a complete description in terms of three sound fundamentals: frequency, amplitude, and time. NEW METHOD Using the techniques of electrophysiology, neural signal processing and medical image processing, a standalone method is created to illustrate the neural processing of three sound fundamentals in one representation. RESULTS The new method comprehensively showed frequency tuning, intensity tuning, time tuning as well as a novel representation of frequency and time dependent intensity coding. It provides most of the necessary parameters that are used to quantify neural response properties, such as minimum threshold (MT), frequency tuning, latency, best frequency (BF), characteristic frequency (CF), bandwidth (BW), etc. COMPARISON WITH EXISTING METHODS: Our method shows neural responses as a function of all three sound fundamentals in a single representation that was not possible in previous methods. It covers many functions of conventional methods and allow extracting novel information such as the intensity coding as the function of the spectrotemporal response area of auditory neurons. CONCLUSION This method can be used as a standalone package to study auditory neural responses and evaluate the performance of different hearing related devices such as cochlear implants and hearing aids in animal models as well as study and compare auditory processing in aged and hearing impaired animal models.
Collapse
Affiliation(s)
- Farhad Qureshi
- Department of Physiology and Pharmacology, University of Calgary, Cumming School of Medicine 3330 Hospital Drive NW, Calgary Alberta, T2N 4N1, Canada.
| | - Jun Yan
- Department of Physiology and Pharmacology, University of Calgary, Cumming School of Medicine 3330 Hospital Drive NW, Calgary Alberta, T2N 4N1, Canada
| |
Collapse
|
5
|
Robust Rate-Place Coding of Resolved Components in Harmonic and Inharmonic Complex Tones in Auditory Midbrain. J Neurosci 2020; 40:2080-2093. [PMID: 31996454 DOI: 10.1523/jneurosci.2337-19.2020] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2019] [Revised: 01/12/2020] [Accepted: 01/16/2020] [Indexed: 11/21/2022] Open
Abstract
Harmonic complex tones (HCTs) commonly occurring in speech and music evoke a strong pitch at their fundamental frequency (F0), especially when they contain harmonics individually resolved by the cochlea. When all frequency components of an HCT are shifted by the same amount, the pitch of the resulting inharmonic tone (IHCT) can also shift, although the envelope repetition rate is unchanged. A rate-place code, whereby resolved harmonics are represented by local maxima in firing rates along the tonotopic axis, has been characterized in the auditory nerve and primary auditory cortex, but little is known about intermediate processing stages. We recorded single-neuron responses to HCT and IHCT with varying F0 and sound level in the inferior colliculus (IC) of unanesthetized rabbits of both sexes. Many neurons showed peaks in firing rate when a low-numbered harmonic aligned with the neuron's characteristic frequency, demonstrating "rate-place" coding. The IC rate-place code was most prevalent for F0 > 800 Hz, was only moderately dependent on sound level over a 40 dB range, and was not sensitive to stimulus harmonicity. A spectral receptive-field model incorporating broadband inhibition better predicted the neural responses than a purely excitatory model, suggesting an enhancement of the rate-place representation by inhibition. Some IC neurons showed facilitation in response to HCT relative to pure tones, similar to cortical "harmonic template neurons" (Feng and Wang, 2017), but to a lesser degree. Our findings shed light on the transformation of rate-place coding of resolved harmonics along the auditory pathway.SIGNIFICANCE STATEMENT Harmonic complex tones are ubiquitous in speech and music and produce strong pitch percepts when they contain frequency components that are individually resolved by the cochlea. Here, we characterize a "rate-place" code for resolved harmonics in the auditory midbrain that is more robust across sound levels than the peripheral rate-place code and insensitive to the harmonic relationships among frequency components. We use a computational model to show that inhibition may play an important role in shaping the rate-place code. Our study fills a major gap in understanding the transformations in neural representations of resolved harmonics along the auditory pathway.
Collapse
|
6
|
Yang Y, Lee J, Kim G. Integration of locomotion and auditory signals in the mouse inferior colliculus. eLife 2020; 9:52228. [PMID: 31987070 PMCID: PMC7004561 DOI: 10.7554/elife.52228] [Citation(s) in RCA: 27] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2019] [Accepted: 01/16/2020] [Indexed: 01/10/2023] Open
Abstract
The inferior colliculus (IC) is the major midbrain auditory integration center, where virtually all ascending auditory inputs converge. Although the IC has been extensively studied for sound processing, little is known about the neural activity of the IC in moving subjects, as frequently happens in natural hearing conditions. Here, by recording neural activity in walking mice, we show that the activity of IC neurons is strongly modulated by locomotion, even in the absence of sound stimuli. Similar modulation was also found in hearing-impaired mice, demonstrating that IC neurons receive non-auditory, locomotion-related neural signals. Sound-evoked activity was attenuated during locomotion, and this attenuation increased frequency selectivity across the neuronal population, while maintaining preferred frequencies. Our results suggest that during behavior, integrating movement-related and auditory information is an essential aspect of sound processing in the IC.
Collapse
Affiliation(s)
- Yoonsun Yang
- Center for Neuroscience Imaging Research (CNIR), Institute for Basic Science (IBS), Suwon, Republic of Korea.,Department of Physiology, School of Medicine, Sungkyunkwan University, Suwon, Republic of Korea
| | - Joonyeol Lee
- Center for Neuroscience Imaging Research (CNIR), Institute for Basic Science (IBS), Suwon, Republic of Korea.,Department of Biomedical Engineering, Sungkyunkwan University, Suwon, Republic of Korea
| | - Gunsoo Kim
- Center for Neuroscience Imaging Research (CNIR), Institute for Basic Science (IBS), Suwon, Republic of Korea.,Department of Biomedical Engineering, Sungkyunkwan University, Suwon, Republic of Korea
| |
Collapse
|
7
|
Lopez Espejo M, Schwartz ZP, David SV. Spectral tuning of adaptation supports coding of sensory context in auditory cortex. PLoS Comput Biol 2019; 15:e1007430. [PMID: 31626624 PMCID: PMC6821137 DOI: 10.1371/journal.pcbi.1007430] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2019] [Revised: 10/30/2019] [Accepted: 09/23/2019] [Indexed: 12/19/2022] Open
Abstract
Perception of vocalizations and other behaviorally relevant sounds requires integrating acoustic information over hundreds of milliseconds. Sound-evoked activity in auditory cortex typically has much shorter latency, but the acoustic context, i.e., sound history, can modulate sound evoked activity over longer periods. Contextual effects are attributed to modulatory phenomena, such as stimulus-specific adaption and contrast gain control. However, an encoding model that links context to natural sound processing has yet to be established. We tested whether a model in which spectrally tuned inputs undergo adaptation mimicking short-term synaptic plasticity (STP) can account for contextual effects during natural sound processing. Single-unit activity was recorded from primary auditory cortex of awake ferrets during presentation of noise with natural temporal dynamics and fully natural sounds. Encoding properties were characterized by a standard linear-nonlinear spectro-temporal receptive field (LN) model and variants that incorporated STP-like adaptation. In the adapting models, STP was applied either globally across all input spectral channels or locally to subsets of channels. For most neurons, models incorporating local STP predicted neural activity as well or better than LN and global STP models. The strength of nonlinear adaptation varied across neurons. Within neurons, adaptation was generally stronger for spectral channels with excitatory than inhibitory gain. Neurons showing improved STP model performance also tended to undergo stimulus-specific adaptation, suggesting a common mechanism for these phenomena. When STP models were compared between passive and active behavior conditions, response gain often changed, but average STP parameters were stable. Thus, spectrally and temporally heterogeneous adaptation, subserved by a mechanism with STP-like dynamics, may support representation of the complex spectro-temporal patterns that comprise natural sounds across wide-ranging sensory contexts.
Collapse
Affiliation(s)
- Mateo Lopez Espejo
- Neuroscience Graduate Program, Oregon Health and Science University, Portland, OR, United States of America
| | - Zachary P. Schwartz
- Neuroscience Graduate Program, Oregon Health and Science University, Portland, OR, United States of America
| | - Stephen V. David
- Oregon Hearing Research Center, Oregon Health and Science University, Portland, OR, United States of America
| |
Collapse
|
8
|
Duets recorded in the wild reveal that interindividually coordinated motor control enables cooperative behavior. Nat Commun 2019; 10:2577. [PMID: 31189912 PMCID: PMC6561963 DOI: 10.1038/s41467-019-10593-3] [Citation(s) in RCA: 35] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2018] [Accepted: 05/21/2019] [Indexed: 01/02/2023] Open
Abstract
Many organisms coordinate rhythmic motor actions with those of a partner to generate cooperative social behavior such as duet singing. The neural mechanisms that enable rhythmic interindividual coordination of motor actions are unknown. Here we investigate the neural basis of vocal duetting behavior by using an approach that enables simultaneous recordings of individual vocalizations and multiunit vocal premotor activity in songbird pairs ranging freely in their natural habitat. We find that in the duet-initiating bird, the onset of the partner’s contribution to the duet triggers a change in rhythm in the periodic neural discharges that are exclusively locked to the initiating bird’s own vocalizations. The resulting interindividually synchronized neural activity pattern elicits vocalizations that perfectly alternate between partners in the ongoing song. We suggest that rhythmic cooperative behavior requires exact interindividual coordination of premotor neural activity, which might be achieved by integration of sensory information originating from the interacting partner. Recording neural activity during coordinated behaviors in controlled environments limits opportunities for understanding natural interactions. Here, the authors record from freely moving duetting birds in their natural habitats to reveal the neural mechanisms of interindividual motor coordination.
Collapse
|
9
|
Siveke I, Lingner A, Ammer JJ, Gleiss SA, Grothe B, Felmy F. A Temporal Filter for Binaural Hearing Is Dynamically Adjusted by Sound Pressure Level. Front Neural Circuits 2019; 13:8. [PMID: 30814933 PMCID: PMC6381077 DOI: 10.3389/fncir.2019.00008] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2018] [Accepted: 01/24/2019] [Indexed: 12/02/2022] Open
Abstract
In natural environments our auditory system is exposed to multiple and diverse signals of fluctuating amplitudes. Therefore, to detect, localize, and single out individual sounds the auditory system has to process and filter spectral and temporal information from both ears. It is known that the overall sound pressure level affects sensory signal transduction and therefore the temporal response pattern of auditory neurons. We hypothesize that the mammalian binaural system utilizes a dynamic mechanism to adjust the temporal filters in neuronal circuits to different overall sound pressure levels. Previous studies proposed an inhibitory mechanism generated by the reciprocally coupled dorsal nuclei of the lateral lemniscus (DNLL) as a temporal neuronal-network filter that suppresses rapid binaural fluctuations. Here we investigated the consequence of different sound levels on this filter during binaural processing. Our in vivo and in vitro electrophysiology in Mongolian gerbils shows that the integration of ascending excitation and contralateral inhibition defines the temporal properties of this inhibitory filter. The time course of this filter depends on the synaptic drive, which is modulated by the overall sound pressure level and N-methyl-D-aspartate receptor (NMDAR) signaling. In psychophysical experiments we tested the temporal perception of humans and show that detection and localization of two subsequent tones changes with the sound pressure level consistent with our physiological results. Together our data support the hypothesis that mammals dynamically adjust their time window for sound detection and localization within the binaural system in a sound level dependent manner.
Collapse
Affiliation(s)
- Ida Siveke
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany.,Institute of Zoology and Neurobiology, Ruhr-Universität Bochum, Bochum, Germany
| | - Andrea Lingner
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany
| | - Julian J Ammer
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany.,Graduate School for Systemic Neurosciences, Ludwig-Maximilians-Universität München, Munich, Germany
| | - Sarah A Gleiss
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany.,Graduate School for Systemic Neurosciences, Ludwig-Maximilians-Universität München, Munich, Germany
| | - Benedikt Grothe
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany.,Graduate School for Systemic Neurosciences, Ludwig-Maximilians-Universität München, Munich, Germany
| | - Felix Felmy
- Department Biology II, Division of Neurobiology, Ludwig-Maximilians-Universität München, Munich, Germany.,Institute of Zoology, University of Veterinary Medicine Hannover, Hannover, Germany
| |
Collapse
|
10
|
Zhang Q, Hu X, Hong B, Zhang B. A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex. PLoS Comput Biol 2019; 15:e1006766. [PMID: 30742609 PMCID: PMC6386396 DOI: 10.1371/journal.pcbi.1006766] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2018] [Revised: 02/22/2019] [Accepted: 12/21/2018] [Indexed: 12/03/2022] Open
Abstract
The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway. When speech enters the ear, it is subjected to a series of processing stages prior to arriving at the auditory cortex. Neurons acting at different processing stages have different response properties. For example, at the auditory midbrain, a neuron may specifically detect the onsets of a frequency component in the speech, whereas in the auditory cortex, a neuron may specifically detect phonetic features. The encoding mechanisms underlying these neuronal functions remain unclear. To address this issue, we designed a hierarchical sparse coding model, inspired by the sparse activity of neurons in the sensory system, to learn features in speech signals. We found that the computing units in different layers exhibited hierarchical extraction of speech sound features, similar to those of neurons in the auditory midbrain and auditory cortex, although the computational principles in these layers were the same. The results suggest that sparse coding and max pooling represent universal computational principles throughout the auditory pathway.
Collapse
Affiliation(s)
- Qingtian Zhang
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
| | - Xiaolin Hu
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
- Center for Brain-Inspired Computing Research (CBICR), Tsinghua University, Beijing, China
- * E-mail:
| | - Bo Hong
- School of Medicine, Tsinghua University, Beijing, China
| | - Bo Zhang
- Department of Computer Science and Technology, Tsinghua University, Beijing, China
- Center for Brain-Inspired Computing Research (CBICR), Tsinghua University, Beijing, China
| |
Collapse
|
11
|
Kessler M, Mamach M, Beutelmann R, Bankstahl JP, Bengel FM, Klump GM, Berding G. Activation in the auditory pathway of the gerbil studied with 18F-FDG PET: effects of anesthesia. Brain Struct Funct 2018; 223:4293-4305. [PMID: 30203305 DOI: 10.1007/s00429-018-1743-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2017] [Accepted: 08/29/2018] [Indexed: 01/20/2023]
Abstract
Here, we present results from an 18F-fluorodeoxyglucose (18F-FDG) positron emission tomography (PET) study in the Mongolian gerbil, a preferred animal model in auditory research. One major issue in preclinical nuclear imaging, as well as in most of the neurophysiological methods investigating auditory processing, is the need of anesthesia. We compared the usability of two types of anesthesia which are frequently employed in electrophysiology, ketamine/xylazine (KX), and fentanyl/midazolam/medetomidine (FMM), for valid measurements of auditory activation with 18F-FDG PET. Gerbils were placed in a sound-shielding box and injected with 18F-FDG. Two acoustic free-field conditions were used: (1) baseline (no stimulation, 25 dB background noise) and (2) 90 dB frequency-modulated tones (FM). After 40 min of 18F-FDG uptake, a 30 min acquisition was performed using a small animal PET/CT system. Blood glucose levels were measured after the uptake phase before scanning. Standardized uptake value ratios for relevant regions were determined after implementing image and volume of interest templates. Scans demonstrated a significantly higher uptake in the inferior colliculus with FM stimulation compared to baseline in awake subjects (+ 12%; p = 0.02) and with FMM anesthesia (+ 13%; p = 0.0012), but not with KX anesthesia. In non-auditory brain regions, no significant difference was detected. Blood glucose levels were significantly higher under KX compared to FMM anesthesia (17.29 ± 0.42 mmol/l vs. 14.30 ± 1.91 mmol/l; p = 0.024). These results suggest that valid 18F-FDG PET measurements of auditory activation comparable to electrophysiology can be obtained from gerbils during opioid-based anesthesia due to its limited effects on interfering blood glucose levels.
Collapse
Affiliation(s)
- M Kessler
- Department of Nuclear Medicine, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany
| | - M Mamach
- Department of Nuclear Medicine, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Department of Medical Physics and Radiation Protection, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany
| | - R Beutelmann
- Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Division for animal Physiology and Behaviour Group, Department for Neuroscience, School of Medicine and Health Sciences, University of Oldenburg, Carl von Ossietzky Str. 9-11, 26129, Oldenburg, Germany
| | - J P Bankstahl
- Department of Nuclear Medicine, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany
| | - F M Bengel
- Department of Nuclear Medicine, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany
| | - G M Klump
- Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Division for animal Physiology and Behaviour Group, Department for Neuroscience, School of Medicine and Health Sciences, University of Oldenburg, Carl von Ossietzky Str. 9-11, 26129, Oldenburg, Germany
| | - Georg Berding
- Department of Nuclear Medicine, Hannover Medical School, Carl-Neuberg-Str. 1, 30625, Hannover, Germany. .,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.
| |
Collapse
|
12
|
Bach JH, Kollmeier B, Anemüller J. Matching Pursuit Analysis of Auditory Receptive Fields' Spectro-Temporal Properties. Front Syst Neurosci 2017; 11:4. [PMID: 28232791 PMCID: PMC5299023 DOI: 10.3389/fnsys.2017.00004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2016] [Accepted: 01/23/2017] [Indexed: 11/13/2022] Open
Abstract
Gabor filters have long been proposed as models for spectro-temporal receptive fields (STRFs), with their specific spectral and temporal rate of modulation qualitatively replicating characteristics of STRF filters estimated from responses to auditory stimuli in physiological data. The present study builds on the Gabor-STRF model by proposing a methodology to quantitatively decompose STRFs into a set of optimally matched Gabor filters through matching pursuit, and by quantitatively evaluating spectral and temporal characteristics of STRFs in terms of the derived optimal Gabor-parameters. To summarize a neuron's spectro-temporal characteristics, we introduce a measure for the “diagonality,” i.e., the extent to which an STRF exhibits spectro-temporal transients which cannot be factorized into a product of a spectral and a temporal modulation. With this methodology, it is shown that approximately half of 52 analyzed zebra finch STRFs can each be well approximated by a single Gabor or a linear combination of two Gabor filters. Moreover, the dominant Gabor functions tend to be oriented either in the spectral or in the temporal direction, with truly “diagonal” Gabor functions rarely being necessary for reconstruction of an STRF's main characteristics. As a toy example for the applicability of STRF and Gabor-STRF filters to auditory detection tasks, we use STRF filters as features in an automatic event detection task and compare them to idealized Gabor filters and mel-frequency cepstral coefficients (MFCCs). STRFs classify a set of six everyday sounds with an accuracy similar to reference Gabor features (94% recognition rate). Spectro-temporal STRF and Gabor features outperform reference spectral MFCCs in quiet and in low noise conditions (down to 0 dB signal to noise ratio).
Collapse
Affiliation(s)
- Jörg-Hendrik Bach
- Medizinische Physik, Universität OldenburgOldenburg, Germany
- Cluster of Excellence Hearing4all, Universität OldenburgOldenburg, Germany
| | - Birger Kollmeier
- Medizinische Physik, Universität OldenburgOldenburg, Germany
- Cluster of Excellence Hearing4all, Universität OldenburgOldenburg, Germany
| | - Jörn Anemüller
- Medizinische Physik, Universität OldenburgOldenburg, Germany
- Cluster of Excellence Hearing4all, Universität OldenburgOldenburg, Germany
- *Correspondence: Jörn Anemüller
| |
Collapse
|
13
|
Yassin L, Pecka M, Kajopoulos J, Gleiss H, Li L, Leibold C, Felmy F. Differences in synaptic and intrinsic properties result in topographic heterogeneity of temporal processing of neurons within the inferior colliculus. Hear Res 2016; 341:79-90. [DOI: 10.1016/j.heares.2016.08.005] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/08/2016] [Revised: 08/15/2016] [Accepted: 08/16/2016] [Indexed: 10/21/2022]
|
14
|
Blackwell JM, Taillefumier TO, Natan RG, Carruthers IM, Magnasco MO, Geffen MN. Stable encoding of sounds over a broad range of statistical parameters in the auditory cortex. Eur J Neurosci 2016; 43:751-64. [PMID: 26663571 PMCID: PMC5021175 DOI: 10.1111/ejn.13144] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2015] [Revised: 11/22/2015] [Accepted: 12/01/2015] [Indexed: 11/29/2022]
Abstract
Natural auditory scenes possess highly structured statistical regularities, which are dictated by the physics of sound production in nature, such as scale‐invariance. We recently identified that natural water sounds exhibit a particular type of scale invariance, in which the temporal modulation within spectral bands scales with the centre frequency of the band. Here, we tested how neurons in the mammalian primary auditory cortex encode sounds that exhibit this property, but differ in their statistical parameters. The stimuli varied in spectro‐temporal density and cyclo‐temporal statistics over several orders of magnitude, corresponding to a range of water‐like percepts, from pattering of rain to a slow stream. We recorded neuronal activity in the primary auditory cortex of awake rats presented with these stimuli. The responses of the majority of individual neurons were selective for a subset of stimuli with specific statistics. However, as a neuronal population, the responses were remarkably stable over large changes in stimulus statistics, exhibiting a similar range in firing rate, response strength, variability and information rate, and only minor variation in receptive field parameters. This pattern of neuronal responses suggests a potentially general principle for cortical encoding of complex acoustic scenes: while individual cortical neurons exhibit selectivity for specific statistical features, a neuronal population preserves a constant response structure across a broad range of statistical parameters.
Collapse
Affiliation(s)
- Jennifer M Blackwell
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Thibaud O Taillefumier
- Center for Physics and Biology, Rockefeller University, New York, NY, USA.,Lewis-Sigler Institute for Integrative Genomics, Princeton University, Princeton, NJ, USA
| | - Ryan G Natan
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Isaac M Carruthers
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Marcelo O Magnasco
- Center for Physics and Biology, Rockefeller University, New York, NY, USA
| | - Maria N Geffen
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA.,Center for Physics and Biology, Rockefeller University, New York, NY, USA
| |
Collapse
|
15
|
Shen L, Zhao L, Hong B. Frequency-specific adaptation and its underlying circuit model in the auditory midbrain. Front Neural Circuits 2015; 9:55. [PMID: 26483641 PMCID: PMC4589587 DOI: 10.3389/fncir.2015.00055] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2015] [Accepted: 09/16/2015] [Indexed: 11/13/2022] Open
Abstract
Receptive fields of sensory neurons are considered to be dynamic and depend on the stimulus history. In the auditory system, evidence of dynamic frequency-receptive fields has been found following stimulus-specific adaptation (SSA). However, the underlying mechanism and circuitry of SSA have not been fully elucidated. Here, we studied how frequency-receptive fields of neurons in rat inferior colliculus (IC) changed when exposed to a biased tone sequence. Pure tone with one specific frequency (adaptor) was presented markedly more often than others. The adapted tuning was compared with the original tuning measured with an unbiased sequence. We found inhomogeneous changes in frequency tuning in IC, exhibiting a center-surround pattern with respect to the neuron's best frequency. Central adaptors elicited strong suppressive and repulsive changes while flank adaptors induced facilitative and attractive changes. Moreover, we proposed a two-layer model of the underlying network, which not only reproduced the adaptive changes in the receptive fields but also predicted novelty responses to oddball sequences. These results suggest that frequency-specific adaptation in auditory midbrain can be accounted for by an adapted frequency channel and its lateral spreading of adaptation, which shed light on the organization of the underlying circuitry.
Collapse
Affiliation(s)
- Li Shen
- Department of Biomedical Engineering, School of Medicine, Tsinghua University Beijing, China
| | - Lingyun Zhao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University Beijing, China
| | - Bo Hong
- Department of Biomedical Engineering, School of Medicine, Tsinghua University Beijing, China
| |
Collapse
|
16
|
Keller CH, Takahashi TT. Spike timing precision changes with spike rate adaptation in the owl's auditory space map. J Neurophysiol 2015; 114:2204-19. [PMID: 26269555 PMCID: PMC4600961 DOI: 10.1152/jn.00442.2015] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2015] [Accepted: 08/07/2015] [Indexed: 11/22/2022] Open
Abstract
Spike rate adaptation (SRA) is a continuing change of responsiveness to ongoing stimuli, which is ubiquitous across species and levels of sensory systems. Under SRA, auditory responses to constant stimuli change over time, relaxing toward a long-term rate often over multiple timescales. With more variable stimuli, SRA causes the dependence of spike rate on sound pressure level to shift toward the mean level of recent stimulus history. A model based on subtractive adaptation (Benda J, Hennig RM. J Comput Neurosci 24: 113-136, 2008) shows that changes in spike rate and level dependence are mechanistically linked. Space-specific neurons in the barn owl's midbrain, when recorded under ketamine-diazepam anesthesia, showed these classical characteristics of SRA, while at the same time exhibiting changes in spike timing precision. Abrupt level increases of sinusoidally amplitude-modulated (SAM) noise initially led to spiking at higher rates with lower temporal precision. Spike rate and precision relaxed toward their long-term values with a time course similar to SRA, results that were also replicated by the subtractive model. Stimuli whose amplitude modulations (AMs) were not synchronous across carrier frequency evoked spikes in response to stimulus envelopes of a particular shape, characterized by the spectrotemporal receptive field (STRF). Again, abrupt stimulus level changes initially disrupted the temporal precision of spiking, which then relaxed along with SRA. We suggest that shifts in latency associated with stimulus level changes may differ between carrier frequency bands and underlie decreased spike precision. Thus SRA is manifest not simply as a change in spike rate but also as a change in the temporal precision of spiking.
Collapse
|
17
|
Fast and robust estimation of spectro-temporal receptive fields using stochastic approximations. J Neurosci Methods 2015; 246:119-33. [PMID: 25744059 DOI: 10.1016/j.jneumeth.2015.02.009] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2014] [Revised: 01/22/2015] [Accepted: 02/11/2015] [Indexed: 11/23/2022]
Abstract
BACKGROUND The receptive field (RF) represents the signal preferences of sensory neurons and is the primary analysis method for understanding sensory coding. While it is essential to estimate a neuron's RF, finding numerical solutions to increasingly complex RF models can become computationally intensive, in particular for high-dimensional stimuli or when many neurons are involved. NEW METHOD Here we propose an optimization scheme based on stochastic approximations that facilitate this task. The basic idea is to derive solutions on a random subset rather than computing the full solution on the available data set. To test this, we applied different optimization schemes based on stochastic gradient descent (SGD) to both the generalized linear model (GLM) and a recently developed classification-based RF estimation approach. RESULTS AND COMPARISON WITH EXISTING METHOD Using simulated and recorded responses, we demonstrate that RF parameter optimization based on state-of-the-art SGD algorithms produces robust estimates of the spectro-temporal receptive field (STRF). Results on recordings from the auditory midbrain demonstrate that stochastic approximations preserve both predictive power and tuning properties of STRFs. A correlation of 0.93 with the STRF derived from the full solution may be obtained in less than 10% of the full solution's estimation time. We also present an on-line algorithm that allows simultaneous monitoring of STRF properties of more than 30 neurons on a single computer. CONCLUSIONS The proposed approach may not only prove helpful for large-scale recordings but also provides a more comprehensive characterization of neural tuning in experiments than standard tuning curves.
Collapse
|
18
|
Genzel D, Hoffmann S, Prosch S, Firzlaff U, Wiegrebe L. Biosonar navigation above water II: exploiting mirror images. J Neurophysiol 2015; 113:1146-55. [DOI: 10.1152/jn.00264.2014] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
As in vision, acoustic signals can be reflected by a smooth surface creating an acoustic mirror image. Water bodies represent the only naturally occurring horizontal and acoustically smooth surfaces. Echolocating bats flying over smooth water bodies encounter echo-acoustic mirror images of objects above the surface. Here, we combined an electrophysiological approach with a behavioral experimental paradigm to investigate whether bats can exploit echo-acoustic mirror images for navigation and how these mirrorlike echo-acoustic cues are encoded in their auditory cortex. In an obstacle-avoidance task where the obstacles could only be detected via their echo-acoustic mirror images, most bats spontaneously exploited these cues for navigation. Sonar ensonifications along the bats' flight path revealed conspicuous changes of the reflection patterns with slightly increased target strengths at relatively long echo delays corresponding to the longer acoustic paths from the mirrored obstacles. Recordings of cortical spatiotemporal response maps (STRMs) describe the tuning of a unit across the dimensions of elevation and time. The majority of cortical single and multiunits showed a special spatiotemporal pattern of excitatory areas in their STRM indicating a preference for echoes with (relative to the setup dimensions) long delays and, interestingly, from low elevations. This neural preference could effectively encode a reflection pattern as it would be perceived by an echolocating bat detecting an object mirrored from below. The current study provides both behavioral and neurophysiological evidence that echo-acoustic mirror images can be exploited by bats for obstacle avoidance. This capability effectively supports echo-acoustic navigation in highly cluttered natural habitats.
Collapse
Affiliation(s)
- Daria Genzel
- Department Biology II, Ludwig-Maximilians-Universität München, Planegg-Martinsried, Germany; and
| | - Susanne Hoffmann
- Department Biology II, Ludwig-Maximilians-Universität München, Planegg-Martinsried, Germany; and
- Chair of Zoology, Technische Universität München, Freising-Weihenstephan, Germany
| | - Selina Prosch
- Department Biology II, Ludwig-Maximilians-Universität München, Planegg-Martinsried, Germany; and
| | - Uwe Firzlaff
- Chair of Zoology, Technische Universität München, Freising-Weihenstephan, Germany
| | - Lutz Wiegrebe
- Department Biology II, Ludwig-Maximilians-Universität München, Planegg-Martinsried, Germany; and
| |
Collapse
|
19
|
Meyer AF, Diepenbrock JP, Ohl FW, Anemüller J. Temporal variability of spectro-temporal receptive fields in the anesthetized auditory cortex. Front Comput Neurosci 2014; 8:165. [PMID: 25566049 PMCID: PMC4274980 DOI: 10.3389/fncom.2014.00165] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2014] [Accepted: 11/30/2014] [Indexed: 11/13/2022] Open
Abstract
Temporal variability of neuronal response characteristics during sensory stimulation is a ubiquitous phenomenon that may reflect processes such as stimulus-driven adaptation, top-down modulation or spontaneous fluctuations. It poses a challenge to functional characterization methods such as the receptive field, since these often assume stationarity. We propose a novel method for estimation of sensory neurons' receptive fields that extends the classic static linear receptive field model to the time-varying case. Here, the long-term estimate of the static receptive field serves as the mean of a probabilistic prior distribution from which the short-term temporally localized receptive field may deviate stochastically with time-varying standard deviation. The derived corresponding generalized linear model permits robust characterization of temporal variability in receptive field structure also for highly non-Gaussian stimulus ensembles. We computed and analyzed short-term auditory spectro-temporal receptive field (STRF) estimates with characteristic temporal resolution 5-30 s based on model simulations and responses from in total 60 single-unit recordings in anesthetized Mongolian gerbil auditory midbrain and cortex. Stimulation was performed with short (100 ms) overlapping frequency-modulated tones. Results demonstrate identification of time-varying STRFs, with obtained predictive model likelihoods exceeding those from baseline static STRF estimation. Quantitative characterization of STRF variability reveals a higher degree thereof in auditory cortex compared to midbrain. Cluster analysis indicates that significant deviations from the long-term static STRF are brief, but reliably estimated. We hypothesize that the observed variability more likely reflects spontaneous or state-dependent internal fluctuations that interact with stimulus-induced processing, rather than experimental or stimulus design.
Collapse
Affiliation(s)
- Arne F Meyer
- Medizinische Physik and Cluster of Excellence Hearing4all, Department of Medical Physics and Acoustics, Carl von Ossietzky University Oldenburg, Germany
| | - Jan-Philipp Diepenbrock
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology Magdeburg, Germany
| | - Frank W Ohl
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology Magdeburg, Germany ; Department of Neuroprosthetics, Institute of Biology, Otto-von-Guericke University Magdeburg, Germany
| | - Jörn Anemüller
- Medizinische Physik and Cluster of Excellence Hearing4all, Department of Medical Physics and Acoustics, Carl von Ossietzky University Oldenburg, Germany
| |
Collapse
|
20
|
Koka K, Tollin DJ. Linear coding of complex sound spectra by discharge rate in neurons of the medial nucleus of the trapezoid body (MNTB) and its inputs. Front Neural Circuits 2014; 8:144. [PMID: 25565971 PMCID: PMC4267272 DOI: 10.3389/fncir.2014.00144] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2014] [Accepted: 11/25/2014] [Indexed: 11/25/2022] Open
Abstract
The interaural level difference (ILD) cue to sound location is first encoded in the lateral superior olive (LSO). ILD sensitivity results because the LSO receives excitatory input from the ipsilateral cochlear nucleus and inhibitory input indirectly from the contralateral cochlear nucleus via glycinergic neurons of the ipsilateral medial nucleus of the trapezoid body (MNTB). It is hypothesized that in order for LSO neurons to encode ILDs, the sound spectra at both ears must be accurately encoded via spike rate by their afferents. This spectral-coding hypothesis has not been directly tested in MNTB, likely because MNTB neurons have been mostly described and studied recently in regards to their abilities to encode temporal aspects of sounds, not spectral. Here, we test the hypothesis that MNTB neurons and their inputs from the cochlear nucleus and auditory nerve code sound spectra via discharge rate. The Random Spectral Shape (RSS) method was used to estimate how the levels of 100-ms duration spectrally stationary stimuli were weighted, both linearly and non-linearly, across a wide band of frequencies. In general, MNTB neurons, and their globular bushy cell inputs, were found to be well-modeled by a linear weighting of spectra demonstrating that the pathways through the MNTB can accurately encode sound spectra including those resulting from the acoustical cues to sound location provided by head-related directional transfer functions (DTFs). Together with the anatomical and biophysical specializations for timing in the MNTB-LSO complex, these mechanisms may allow ILDs to be computed for complex stimuli with rapid spectrotemporally-modulated envelopes such as speech and animal vocalizations and moving sound sources.
Collapse
Affiliation(s)
- Kanthaiah Koka
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| | - Daniel J Tollin
- Department of Physiology and Biophysics, University of Colorado School of Medicine Aurora, CO, USA
| |
Collapse
|
21
|
A new and fast characterization of multiple encoding properties of auditory neurons. Brain Topogr 2014; 28:379-400. [PMID: 24869676 DOI: 10.1007/s10548-014-0375-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2013] [Accepted: 05/07/2014] [Indexed: 10/25/2022]
Abstract
The functional properties of auditory cortex neurons are most often investigated separately, through spectrotemporal receptive fields (STRFs) for the frequency tuning and the use of frequency sweeps sounds for selectivity to velocity and direction. In fact, auditory neurons are sensitive to a multidimensional space of acoustic parameters where spectral, temporal and spatial dimensions interact. We designed a multi-parameter stimulus, the random double sweep (RDS), composed of two uncorrelated random sweeps, which gives an easy, fast and simultaneous access to frequency tuning as well as frequency modulation sweep direction and velocity selectivity, frequency interactions and temporal properties of neurons. Reverse correlation techniques applied to recordings from the primary auditory cortex of guinea pigs and rats in response to RDS stimulation revealed the variety of temporal dynamics of acoustic patterns evoking an enhanced or suppressed firing rate. Group results on these two species revealed less frequent suppression areas in frequency tuning STRFs, the absence of downward sweep selectivity, and lower phase locking abilities in the auditory cortex of rats compared to guinea pigs.
Collapse
|
22
|
Fontaine B, MacLeod KM, Lubejko ST, Steinberg LJ, Köppl C, Peña JL. Emergence of band-pass filtering through adaptive spiking in the owl's cochlear nucleus. J Neurophysiol 2014; 112:430-45. [PMID: 24790170 DOI: 10.1152/jn.00132.2014] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
In the visual, auditory, and electrosensory modalities, stimuli are defined by first- and second-order attributes. The fast time-pressure signal of a sound, a first-order attribute, is important, for instance, in sound localization and pitch perception, while its slow amplitude-modulated envelope, a second-order attribute, can be used for sound recognition. Ascending the auditory pathway from ear to midbrain, neurons increasingly show a preference for the envelope and are most sensitive to particular envelope modulation frequencies, a tuning considered important for encoding sound identity. The level at which this tuning property emerges along the pathway varies across species, and the mechanism of how this occurs is a matter of debate. In this paper, we target the transition between auditory nerve fibers and the cochlear nucleus angularis (NA). While the owl's auditory nerve fibers simultaneously encode the fast and slow attributes of a sound, one synapse further, NA neurons encode the envelope more efficiently than the auditory nerve. Using in vivo and in vitro electrophysiology and computational analysis, we show that a single-cell mechanism inducing spike threshold adaptation can explain the difference in neural filtering between the two areas. We show that spike threshold adaptation can explain the increased selectivity to modulation frequency, as input level increases in NA. These results demonstrate that a spike generation nonlinearity can modulate the tuning to second-order stimulus features, without invoking network or synaptic mechanisms.
Collapse
Affiliation(s)
- Bertrand Fontaine
- Dominick Purpura Department of Neuroscience, Albert Einstein College of Medicine, Bronx, New York;
| | - Katrina M MacLeod
- Department of Biology, Neuroscience and Cognitive Science Program, University of Maryland, College Park, Maryland; and
| | - Susan T Lubejko
- Department of Biology, Neuroscience and Cognitive Science Program, University of Maryland, College Park, Maryland; and
| | - Louisa J Steinberg
- Dominick Purpura Department of Neuroscience, Albert Einstein College of Medicine, Bronx, New York
| | - Christine Köppl
- Cluster of Excellence "Hearing4all" and Research Center Neurosensory Science and Department of Neuroscience School of Medicine and Health Science, Carl von Ossietzky University, Oldenburg, Germany
| | - Jose L Peña
- Dominick Purpura Department of Neuroscience, Albert Einstein College of Medicine, Bronx, New York
| |
Collapse
|
23
|
Meyer AF, Diepenbrock JP, Happel MFK, Ohl FW, Anemüller J. Discriminative learning of receptive fields from responses to non-Gaussian stimulus ensembles. PLoS One 2014; 9:e93062. [PMID: 24699631 PMCID: PMC3974709 DOI: 10.1371/journal.pone.0093062] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2013] [Accepted: 02/28/2014] [Indexed: 11/19/2022] Open
Abstract
Analysis of sensory neurons' processing characteristics requires simultaneous measurement of presented stimuli and concurrent spike responses. The functional transformation from high-dimensional stimulus space to the binary space of spike and non-spike responses is commonly described with linear-nonlinear models, whose linear filter component describes the neuron's receptive field. From a machine learning perspective, this corresponds to the binary classification problem of discriminating spike-eliciting from non-spike-eliciting stimulus examples. The classification-based receptive field (CbRF) estimation method proposed here adapts a linear large-margin classifier to optimally predict experimental stimulus-response data and subsequently interprets learned classifier weights as the neuron's receptive field filter. Computational learning theory provides a theoretical framework for learning from data and guarantees optimality in the sense that the risk of erroneously assigning a spike-eliciting stimulus example to the non-spike class (and vice versa) is minimized. Efficacy of the CbRF method is validated with simulations and for auditory spectro-temporal receptive field (STRF) estimation from experimental recordings in the auditory midbrain of Mongolian gerbils. Acoustic stimulation is performed with frequency-modulated tone complexes that mimic properties of natural stimuli, specifically non-Gaussian amplitude distribution and higher-order correlations. Results demonstrate that the proposed approach successfully identifies correct underlying STRFs, even in cases where second-order methods based on the spike-triggered average (STA) do not. Applied to small data samples, the method is shown to converge on smaller amounts of experimental recordings and with lower estimation variance than the generalized linear model and recent information theoretic methods. Thus, CbRF estimation may prove useful for investigation of neuronal processes in response to natural stimuli and in settings where rapid adaptation is induced by experimental design.
Collapse
Affiliation(s)
- Arne F. Meyer
- Department of Medical Physics and Acoustics and Cluster of Excellence ''Hearing4all'', University of Oldenburg, Oldenburg, Germany
- * E-mail:
| | - Jan-Philipp Diepenbrock
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
| | - Max F. K. Happel
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
- Department of Neuroprosthetics, Institute of Biology, Otto-von-Guericke University, Magdeburg, Germany
| | - Frank W. Ohl
- Department of Systems Physiology of Learning, Leibniz Institute for Neurobiology, Magdeburg, Germany
- Department of Neuroprosthetics, Institute of Biology, Otto-von-Guericke University, Magdeburg, Germany
| | - Jörn Anemüller
- Department of Medical Physics and Acoustics and Cluster of Excellence ''Hearing4all'', University of Oldenburg, Oldenburg, Germany
| |
Collapse
|
24
|
Schafer PB, Jin DZ. Noise-Robust Speech Recognition Through Auditory Feature Detection and Spike Sequence Decoding. Neural Comput 2014; 26:523-56. [DOI: 10.1162/neco_a_00557] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Speech recognition in noisy conditions is a major challenge for computer systems, but the human brain performs it routinely and accurately. Automatic speech recognition (ASR) systems that are inspired by neuroscience can potentially bridge the performance gap between humans and machines. We present a system for noise-robust isolated word recognition that works by decoding sequences of spikes from a population of simulated auditory feature-detecting neurons. Each neuron is trained to respond selectively to a brief spectrotemporal pattern, or feature, drawn from the simulated auditory nerve response to speech. The neural population conveys the time-dependent structure of a sound by its sequence of spikes. We compare two methods for decoding the spike sequences—one using a hidden Markov model–based recognizer, the other using a novel template-based recognition scheme. In the latter case, words are recognized by comparing their spike sequences to template sequences obtained from clean training data, using a similarity measure based on the length of the longest common sub-sequence. Using isolated spoken digits from the AURORA-2 database, we show that our combined system outperforms a state-of-the-art robust speech recognizer at low signal-to-noise ratios. Both the spike-based encoding scheme and the template-based decoding offer gains in noise robustness over traditional speech recognition methods. Our system highlights potential advantages of spike-based acoustic coding and provides a biologically motivated framework for robust ASR development.
Collapse
Affiliation(s)
- Phillip B. Schafer
- Department of Physics and Center for Neural Engineering, The Pennsylvania State University, University Park, PA 16802, U.S.A
| | - Dezhe Z. Jin
- Department of Physics and Center for Neural Engineering, The Pennsylvania State University, University Park, PA 16802, U.S.A
| |
Collapse
|
25
|
Grimsley CA, Sanchez JT, Sivaramakrishnan S. Midbrain local circuits shape sound intensity codes. Front Neural Circuits 2013; 7:174. [PMID: 24198763 PMCID: PMC3812908 DOI: 10.3389/fncir.2013.00174] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2013] [Accepted: 10/09/2013] [Indexed: 12/28/2022] Open
Abstract
Hierarchical processing of sensory information requires interaction at multiple levels along the peripheral to central pathway. Recent evidence suggests that interaction between driving and modulating components can shape both top down and bottom up processing of sensory information. Here we show that a component inherited from extrinsic sources combines with local components to code sound intensity. By applying high concentrations of divalent cations to neurons in the nucleus of the inferior colliculus in the auditory midbrain, we show that as sound intensity increases, the source of synaptic efficacy changes from inherited inputs to local circuits. In neurons with a wide dynamic range response to intensity, inherited inputs increase firing rates at low sound intensities but saturate at mid-to-high intensities. Local circuits activate at high sound intensities and widen dynamic range by continuously increasing their output gain with intensity. Inherited inputs are necessary and sufficient to evoke tuned responses, however local circuits change peak output. Push–pull driving inhibition and excitation create net excitatory drive to intensity-variant neurons and tune neurons to intensity. Our results reveal that dynamic range and tuning re-emerge in the auditory midbrain through local circuits that are themselves variable or tuned.
Collapse
Affiliation(s)
- Calum Alex Grimsley
- Department of Anatomy and Neurobiology, Northeast Ohio Medical University Rootstown, OH, USA
| | | | | |
Collapse
|
26
|
Conserved mechanisms of vocalization coding in mammalian and songbird auditory midbrain. Hear Res 2013; 305:45-56. [PMID: 23726970 DOI: 10.1016/j.heares.2013.05.005] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/17/2012] [Revised: 03/23/2013] [Accepted: 05/11/2013] [Indexed: 11/23/2022]
Abstract
The ubiquity of social vocalizations among animals provides the opportunity to identify conserved mechanisms of auditory processing that subserve communication. Identifying auditory coding properties that are shared across vocal communicators will provide insight into how human auditory processing leads to speech perception. Here, we compare auditory response properties and neural coding of social vocalizations in auditory midbrain neurons of mammalian and avian vocal communicators. The auditory midbrain is a nexus of auditory processing because it receives and integrates information from multiple parallel pathways and provides the ascending auditory input to the thalamus. The auditory midbrain is also the first region in the ascending auditory system where neurons show complex tuning properties that are correlated with the acoustics of social vocalizations. Single unit studies in mice, bats and zebra finches reveal shared principles of auditory coding including tonotopy, excitatory and inhibitory interactions that shape responses to vocal signals, nonlinear response properties that are important for auditory coding of social vocalizations and modulation tuning. Additionally, single neuron responses in the mouse and songbird midbrain are reliable, selective for specific syllables, and rely on spike timing for neural discrimination of distinct vocalizations. We propose that future research on auditory coding of vocalizations in mouse and songbird midbrain neurons adopt similar experimental and analytical approaches so that conserved principles of vocalization coding may be distinguished from those that are specialized for each species. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives".
Collapse
|
27
|
Yu JJ, Young ED. Frequency response areas in the inferior colliculus: nonlinearity and binaural interaction. Front Neural Circuits 2013; 7:90. [PMID: 23675323 PMCID: PMC3650518 DOI: 10.3389/fncir.2013.00090] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2013] [Accepted: 04/22/2013] [Indexed: 11/13/2022] Open
Abstract
The tuning, binaural properties, and encoding characteristics of neurons in the central nucleus of the inferior colliculus (CNIC) were investigated to shed light on nonlinearities in the responses of these neurons. Results were analyzed for three types of neurons (I, O, and V) in the CNIC of decerebrate cats. Rate responses to binaural stimuli were characterized using a 1st- plus 2nd-order spectral integration model. Parameters of the model were derived using broadband stimuli with random spectral shapes (RSS). This method revealed four characteristics of CNIC neurons: (1) Tuning curves derived from broadband stimuli have fixed (i. e., level tolerant) bandwidths across a 50-60 dB range of sound levels; (2) 1st-order contralateral weights (particularly for type I and O neurons) were usually larger in magnitude than corresponding ipsilateral weights; (3) contralateral weights were more important than ipsilateral weights when using the model to predict responses to untrained noise stimuli; and (4) 2nd-order weight functions demonstrate frequency selectivity different from that of 1st-order weight functions. Furthermore, while the inclusion of 2nd-order terms in the model usually improved response predictions related to untrained RSS stimuli, they had limited impact on predictions related to other forms of filtered broadband noise [e. g., virtual-space stimuli (VS)]. The accuracy of the predictions varied considerably by response type. Predictions were most accurate for I neurons, and less accurate for O and V neurons, except at the lowest stimulus levels. These differences in prediction performance support the idea that type I, O, and V neurons encode different aspects of the stimulus: while type I neurons are most capable of producing linear representations of spectral shape, type O and V neurons may encode spectral features or temporal stimulus properties in a manner not easily explained with the low-order model. Supported by NIH grant DC00115.
Collapse
Affiliation(s)
| | - Eric D. Young
- Center for Hearing and Balance, Department of Biomedical Engineering, Johns Hopkins UniversityBaltimore, MD, USA
| |
Collapse
|
28
|
Chang TR, Chiu TW, Sun X, Poon PWF. Modeling complex responses of FM-sensitive cells in the auditory midbrain using a committee machine. Brain Res 2013; 1536:44-52. [PMID: 23665390 DOI: 10.1016/j.brainres.2013.04.058] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2012] [Revised: 04/30/2013] [Accepted: 04/30/2013] [Indexed: 11/26/2022]
Abstract
Frequency modulation (FM) is an important building block of complex sounds that include speech signals. Exploring the neural mechanisms of FM coding with computer modeling could help understand how speech sounds are processed in the brain. Here, we modeled the single unit responses of auditory neurons recorded from the midbrain of anesthetized rats. These neurons displayed spectral temporal receptive fields (STRFs) that had multiple-trigger features, and were more complex than those with single-trigger features. Their responses have not been modeled satisfactorily with simple artificial neural networks, unlike neurons with simple-trigger features. To improve model performance, here we tested an approach with the committee machine. For a given neuron, the peri-stimulus time histogram (PSTH) was first generated in response to a repeated random FM tone, and peaks in the PSTH were segregated into groups based on the similarity of their pre-spike FM trigger features. Each group was then modeled using an artificial neural network with simple architecture, and, when necessary, by increasing the number of neurons in the hidden layer. After initial training, the artificial neural networks with their optimized weighting coefficients were pooled into a committee machine for training. Finally, the model performance was tested by prediction of the response of the same cell to a novel FM tone. The results showed improvement over simple artificial neural networks, supporting that trigger-feature-based modeling can be extended to cells with complex responses. This article is part of a Special Issue entitled Neural Coding 2012. This article is part of a Special Issue entitled Neural Coding 2012.
Collapse
Affiliation(s)
- T R Chang
- Department of Computer Science and Information Engineering, Southern Taiwan University of Science and Technology, Tainan, Taiwan.
| | | | | | | |
Collapse
|
29
|
Linear processing of interaural level difference underlies spatial tuning in the nucleus of the brachium of the inferior colliculus. J Neurosci 2013; 33:3891-904. [PMID: 23447600 DOI: 10.1523/jneurosci.3437-12.2013] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
The spatial location of sounds is an important aspect of auditory perception, but the ways in which space is represented are not fully understood. No space map has been found within the primary auditory pathway. However, a space map has been found in the nucleus of the brachium of the inferior colliculus (BIN), which provides a major auditory projection to the superior colliculus. We measured the spectral processing underlying auditory spatial tuning in the BIN of unanesthetized marmoset monkeys. Because neurons in the BIN respond poorly to tones and are broadly tuned, we used a broadband stimulus with random spectral shapes (RSSs) from which both spatial receptive fields and frequency sensitivity can be derived. Responses to virtual space (VS) stimuli, based on the animal's own ear acoustics, were compared with the predictions of a weight-function model of responses to the RSS stimuli. First-order (linear) weight functions had broad spectral tuning (approximately three octaves) and were excitatory in the contralateral ear, inhibitory in the ipsilateral ear, and biased toward high frequencies. Responses to interaural time differences and spectral cues were relatively weak. In cross-validation tests, the first-order RSS model accurately predicted the measured VS tuning curves in the majority of neurons, but was inaccurate in 25% of neurons. In some cases, second-order weighting functions led to significant improvements. Finally, we found a significant correlation between the degree of binaural weight asymmetry and the best azimuth. Overall, the results suggest that linear processing of interaural level difference underlies spatial tuning in the BIN.
Collapse
|
30
|
Recognition of non-harmonic natural sounds by small mammals using competitive training. PLoS One 2012; 7:e51318. [PMID: 23251497 PMCID: PMC3519678 DOI: 10.1371/journal.pone.0051318] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2012] [Accepted: 10/31/2012] [Indexed: 11/19/2022] Open
Abstract
Animals recognize biologically relevant sounds, such as the non-harmonic sounds made by some predators, and respond with adaptive behaviors, such as escaping. To clarify which acoustic parameters are used for identifying non-harmonic, noise-like, broadband sounds, guinea pigs were conditioned to a natural target sound by introducing a novel training procedure in which 2 or 3 guinea pigs in a group competed for food. A set of distinct behavioral reactions was reliably induced almost exclusively to the target sound in a 2-week operant training. When fully conditioned, individual animals were separately tested for recognition of a set of target-like sounds that had been modified from the target sound, with spectral ranges eliminated or with fine or coarse temporal structures altered. The results show that guinea pigs are able to identify the noise-like non-harmonic natural sounds by relying on gross spectral compositions and/or fine temporal structures, just as birds are thought to do in the recognition of harmonic birdsongs. These findings are discussed with regard to similarities and dissimilarities to harmonic sound recognition. The results suggest that similar but not identical processing that requires different time scales might be used to recognize harmonic and non-harmonic sounds, at least in small mammals.
Collapse
|
31
|
Laudanski J, Edeline JM, Huetz C. Differences between spectro-temporal receptive fields derived from artificial and natural stimuli in the auditory cortex. PLoS One 2012; 7:e50539. [PMID: 23209771 PMCID: PMC3507792 DOI: 10.1371/journal.pone.0050539] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2012] [Accepted: 10/25/2012] [Indexed: 11/25/2022] Open
Abstract
Spectro-temporal properties of auditory cortex neurons have been extensively studied with artificial sounds but it is still unclear whether they help in understanding neuronal responses to communication sounds. Here, we directly compared spectro-temporal receptive fields (STRFs) obtained from the same neurons using both artificial stimuli (dynamic moving ripples, DMRs) and natural stimuli (conspecific vocalizations) that were matched in terms of spectral content, average power and modulation spectrum. On a population of auditory cortex neurons exhibiting reliable tuning curves when tested with pure tones, significant STRFs were obtained for 62% of the cells with vocalizations and 68% with DMR. However, for many cells with significant vocalization-derived STRFs (STRFvoc) and DMR-derived STRFs (STRFdmr), the BF, latency, bandwidth and global STRFs shape differed more than what would be predicted by spiking responses simulated by a linear model based on a non-homogenous Poisson process. Moreover STRFvoc predicted neural responses to vocalizations more accurately than STRFdmr predicted neural response to DMRs, despite similar spike-timing reliability for both sets of stimuli. Cortical bursts, which potentially introduce nonlinearities in evoked responses, did not explain the differences between STRFvoc and STRFdmr. Altogether, these results suggest that the nonlinearity of auditory cortical responses makes it difficult to predict responses to communication sounds from STRFs computed from artificial stimuli.
Collapse
Affiliation(s)
- Jonathan Laudanski
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| | - Jean-Marc Edeline
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
- * E-mail:
| | - Chloé Huetz
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| |
Collapse
|
32
|
Carlson NL, Ming VL, DeWeese MR. Sparse codes for speech predict spectrotemporal receptive fields in the inferior colliculus. PLoS Comput Biol 2012; 8:e1002594. [PMID: 22807665 PMCID: PMC3395612 DOI: 10.1371/journal.pcbi.1002594] [Citation(s) in RCA: 41] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2011] [Accepted: 05/18/2012] [Indexed: 12/03/2022] Open
Abstract
We have developed a sparse mathematical representation of speech that minimizes the number of active model neurons needed to represent typical speech sounds. The model learns several well-known acoustic features of speech such as harmonic stacks, formants, onsets and terminations, but we also find more exotic structures in the spectrogram representation of sound such as localized checkerboard patterns and frequency-modulated excitatory subregions flanked by suppressive sidebands. Moreover, several of these novel features resemble neuronal receptive fields reported in the Inferior Colliculus (IC), as well as auditory thalamus and cortex, and our model neurons exhibit the same tradeoff in spectrotemporal resolution as has been observed in IC. To our knowledge, this is the first demonstration that receptive fields of neurons in the ascending mammalian auditory pathway beyond the auditory nerve can be predicted based on coding principles and the statistical properties of recorded sounds.
Collapse
Affiliation(s)
- Nicole L. Carlson
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
- Department of Physics, University of California, Berkeley, California, United States of America
| | - Vivienne L. Ming
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
| | - Michael Robert DeWeese
- Redwood Center for Theoretical Neuroscience, University of California, Berkeley, California, United States of America
- Department of Physics, University of California, Berkeley, California, United States of America
- Helen Wills Neuroscience Institute, University of California, Berkeley, California, United States of America
| |
Collapse
|
33
|
Atencio CA, Sharpee TO, Schreiner CE. Receptive field dimensionality increases from the auditory midbrain to cortex. J Neurophysiol 2012; 107:2594-603. [PMID: 22323634 DOI: 10.1152/jn.01025.2011] [Citation(s) in RCA: 45] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
In the primary auditory cortex, spectrotemporal receptive fields (STRFs) are composed of multiple independent components that capture the processing of disparate stimulus aspects by any given neuron. The origin of these multidimensional stimulus filters in the central auditory system is unknown. To determine whether multicomponent STRFs emerge prior to the forebrain, we recorded from single neurons in the main obligatory station of the auditory midbrain, the inferior colliculus. By comparing results of different spike-triggered techniques, we found that the neural responses in the inferior colliculus can be accounted for by a single stimulus filter. This was observed for all temporal response patterns, from strongly phasic to tonic. Our results reveal that spectrotemporal stimulus encoding undergoes a fundamental transformation along the auditory neuraxis, with the emergence of multidimensional receptive fields beyond the auditory midbrain.
Collapse
Affiliation(s)
- Craig A Atencio
- The UCSF Center for Integrative Neuroscience, Department of Otolaryngology-Headand Neck Surgery, University of California, San Francisco, CA, USA.
| | | | | |
Collapse
|
34
|
Razak KA. Mechanisms underlying intensity-dependent changes in cortical selectivity for frequency-modulated sweeps. J Neurophysiol 2012; 107:2202-11. [PMID: 22279192 DOI: 10.1152/jn.00922.2011] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Frequency-modulated (FM) sweeps are common components of species-specific vocalizations. The intensity of FM sweeps can cover a wide range in the natural environment, but whether intensity affects neural selectivity for FM sweeps is unclear. Bats, such as the pallid bat, which use FM sweeps for echolocation, are suited to address this issue, because the intensity of echoes will vary with target distance. In this study, FM sweep rate selectivity of pallid bat auditory cortex neurons was measured using downward sweeps at different intensities. Neurons became more selective for FM sweep rates present in the bat's echolocation calls as intensity increased. Increased selectivity resulted from stronger inhibition of responses to slower sweep rates. The timing and bandwidth of inhibition generated by frequencies on the high side of the excitatory tuning curve [sideband high-frequency inhibition (HFI)] shape rate selectivity in cortical neurons in the pallid bat. To determine whether intensity-dependent changes in FM rate selectivity were due to altered inhibition, the timing and bandwidth of HFI were quantified at multiple intensities using the two-tone inhibition paradigm. HFI arrived faster relative to excitation as sound intensity increased. The bandwidth of HFI also increased with intensity. The changes in HFI predicted intensity-dependent changes in FM rate selectivity. These data suggest that neural selectivity for a sweep parameter is not static but shifts with intensity due to changes in properties of sideband inhibition.
Collapse
Affiliation(s)
- K A Razak
- Dept. of Psychology, Graduate Neuroscience Program, Univ. of California, Riverside, CA 92521, USA.
| |
Collapse
|
35
|
Selectivity for spectral motion as a neural computation for encoding natural communication signals in bat inferior colliculus. J Neurosci 2012; 31:16529-40. [PMID: 22090479 DOI: 10.1523/jneurosci.1306-11.2011] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
This study examines the neural computations performed by neurons in the auditory system to be selective for the direction and velocity of signals sweeping upward or downward in frequency, termed spectral motion. We show that neurons in the auditory midbrain of Mexican free-tailed bats encode multiple spectrotemporal features of natural communication sounds. These features to which each neuron is tuned are nonlinearly combined to produce selectivity for spectral motion cues present in their conspecific calls, such as direction and velocity. We find that the neural computations resulting in selectivity for spectral motion are analogous to models of motion selectivity studied in vision. Our analysis revealed that auditory neurons in the inferior colliculus (IC) are avoiding spectrotemporal modulations that are redundant across different bat communication signals and are specifically tuned for modulations that distinguish each call from another by their frequency-modulated direction and velocity, suggesting that spectral motion is the neural computation through which IC neurons are encoding specific features of conspecific vocalizations.
Collapse
|
36
|
Extra-classical tuning predicts stimulus-dependent receptive fields in auditory neurons. J Neurosci 2011; 31:11867-78. [PMID: 21849547 DOI: 10.1523/jneurosci.5790-10.2011] [Citation(s) in RCA: 32] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
The receptive fields of many sensory neurons are sensitive to statistical differences among classes of complex stimuli. For example, excitatory spectral bandwidths of midbrain auditory neurons and the spatial extent of cortical visual neurons differ during the processing of natural stimuli compared to the processing of artificial stimuli. Experimentally characterizing neuronal nonlinearities that contribute to stimulus-dependent receptive fields is important for understanding how neurons respond to different stimulus classes in multiple sensory modalities. Here we show that in the zebra finch, many auditory midbrain neurons have extra-classical receptive fields, consisting of sideband excitation and sideband inhibition. We also show that the presence, degree, and asymmetry of stimulus-dependent receptive fields during the processing of complex sounds are predicted by the presence, valence, and asymmetry of extra-classical tuning. Neurons for which excitatory bandwidth expands during the processing of song have extra-classical excitation. Neurons for which frequency tuning is static and for which excitatory bandwidth contracts during the processing of song have extra-classical inhibition. Simulation experiments further demonstrate that stimulus-dependent receptive fields can arise from extra-classical tuning with a static spike threshold nonlinearity. These findings demonstrate that a common neuronal nonlinearity can account for the stimulus dependence of receptive fields estimated from the responses of auditory neurons to stimuli with natural and non-natural statistics.
Collapse
|
37
|
Sharpee TO, Nagel KI, Doupe AJ. Two-dimensional adaptation in the auditory forebrain. J Neurophysiol 2011; 106:1841-61. [PMID: 21753019 PMCID: PMC3296429 DOI: 10.1152/jn.00905.2010] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2010] [Accepted: 07/07/2011] [Indexed: 11/22/2022] Open
Abstract
Sensory neurons exhibit two universal properties: sensitivity to multiple stimulus dimensions, and adaptation to stimulus statistics. How adaptation affects encoding along primary dimensions is well characterized for most sensory pathways, but if and how it affects secondary dimensions is less clear. We studied these effects for neurons in the avian equivalent of primary auditory cortex, responding to temporally modulated sounds. We showed that the firing rate of single neurons in field L was affected by at least two components of the time-varying sound log-amplitude. When overall sound amplitude was low, neural responses were based on nonlinear combinations of the mean log-amplitude and its rate of change (first time differential). At high mean sound amplitude, the two relevant stimulus features became the first and second time derivatives of the sound log-amplitude. Thus a strikingly systematic relationship between dimensions was conserved across changes in stimulus intensity, whereby one of the relevant dimensions approximated the time differential of the other dimension. In contrast to stimulus mean, increases in stimulus variance did not change relevant dimensions, but selectively increased the contribution of the second dimension to neural firing, illustrating a new adaptive behavior enabled by multidimensional encoding. Finally, we demonstrated theoretically that inclusion of time differentials as additional stimulus features, as seen so prominently in the single-neuron responses studied here, is a useful strategy for encoding naturalistic stimuli, because it can lower the necessary sampling rate while maintaining the robustness of stimulus reconstruction to correlated noise.
Collapse
Affiliation(s)
- Tatyana O Sharpee
- The Crick-Jacobs Center for Theoretical and Computational Biology, Computational Neurobiology Laboratory, The Salk Institute for Biological Studies, and the Center for Theoretical Biological Physics, University of California, San Diego, La Jolla, CA, USA.
| | | | | |
Collapse
|
38
|
Chang TR, Chiu TW, Sun X, Poon PWF. Modeling frequency modulated responses of midbrain auditory neurons based on trigger features and artificial neural networks. Brain Res 2011; 1434:90-101. [PMID: 22035565 DOI: 10.1016/j.brainres.2011.09.042] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2011] [Revised: 09/20/2011] [Accepted: 09/21/2011] [Indexed: 11/25/2022]
Abstract
Frequency modulation (FM) is an important building block of communication signals for animals and human. Attempts to predict the response of central neurons to FM sounds have not been very successful, though achieving successful results could bring insights regarding the underlying neural mechanisms. Here we proposed a new method to predict responses of FM-sensitive neurons in the auditory midbrain. First we recorded single unit responses in anesthetized rats using a random FM tone to construct their spectro-temporal receptive fields (STRFs). Training of neurons in the artificial neural network to respond to a second random FM tone was based on the temporal information derived from the STRF. Specifically, the time window covered by the presumed trigger feature and its delay time to spike occurrence were used to train a finite impulse response neural network (FIRNN) to respond to this random FM. Finally we tested the model performance in predicting the response to another similar FM stimuli (third random FM tone). We found good performance in predicting the time of responses if not also the response magnitudes. Furthermore, the weighting function of the FIRNN showed temporal 'bumps' suggesting temporal integration of synaptic inputs from different frequency laminae. This article is part of a Special Issue entitled: Neural Coding.
Collapse
Affiliation(s)
- T R Chang
- Dept. of Computer Sciences and Information Engineering, Southern Taiwan University, Tainan, Taiwan.
| | | | | | | |
Collapse
|
39
|
Zhao L, Zhaoping L. Understanding auditory spectro-temporal receptive fields and their changes with input statistics by efficient coding principles. PLoS Comput Biol 2011; 7:e1002123. [PMID: 21887121 PMCID: PMC3158037 DOI: 10.1371/journal.pcbi.1002123] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2010] [Accepted: 05/31/2011] [Indexed: 11/18/2022] Open
Abstract
Spectro-temporal receptive fields (STRFs) have been widely used as linear approximations to the signal transform from sound spectrograms to neural responses along the auditory pathway. Their dependence on statistical attributes of the stimuli, such as sound intensity, is usually explained by nonlinear mechanisms and models. Here, we apply an efficient coding principle which has been successfully used to understand receptive fields in early stages of visual processing, in order to provide a computational understanding of the STRFs. According to this principle, STRFs result from an optimal tradeoff between maximizing the sensory information the brain receives, and minimizing the cost of the neural activities required to represent and transmit this information. Both terms depend on the statistical properties of the sensory inputs and the noise that corrupts them. The STRFs should therefore depend on the input power spectrum and the signal-to-noise ratio, which is assumed to increase with input intensity. We analytically derive the optimal STRFs when signal and noise are approximated as Gaussians. Under the constraint that they should be spectro-temporally local, the STRFs are predicted to adapt from being band-pass to low-pass filters as the input intensity reduces, or the input correlation becomes longer range in sound frequency or time. These predictions qualitatively match physiological observations. Our prediction as to how the STRFs should be determined by the input power spectrum could readily be tested, since this spectrum depends on the stimulus ensemble. The potentials and limitations of the efficient coding principle are discussed.
Collapse
Affiliation(s)
- Lingyun Zhao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, P.R. China
| | - Li Zhaoping
- Department of Computer Science, University College London, London, United Kingdom
- * E-mail:
| |
Collapse
|
40
|
Sharpee TO, Atencio CA, Schreiner CE. Hierarchical representations in the auditory cortex. Curr Opin Neurobiol 2011; 21:761-7. [PMID: 21704508 DOI: 10.1016/j.conb.2011.05.027] [Citation(s) in RCA: 65] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2011] [Revised: 05/18/2011] [Accepted: 05/20/2011] [Indexed: 11/20/2022]
Abstract
Understanding the neural mechanisms of invariant object recognition remains one of the major unsolved problems in neuroscience. A common solution that is thought to be employed by diverse sensory systems is to create hierarchical representations of increasing complexity and tolerance. However, in the mammalian auditory system many aspects of this hierarchical organization remain undiscovered, including the prominent classes of high-level representations (that would be analogous to face selectivity in the visual system or selectivity to bird's own song in the bird) and the dominant types of invariant transformations. Here we review the recent progress that begins to probe the hierarchy of auditory representations, and the computational approaches that can be helpful in achieving this feat.
Collapse
Affiliation(s)
- Tatyana O Sharpee
- The Salk Institute for Biological Studies, La Jolla CA 92037, United States.
| | | | | |
Collapse
|
41
|
The effects of interaural time difference and intensity on the coding of low-frequency sounds in the mammalian midbrain. J Neurosci 2011; 31:3821-7. [PMID: 21389237 DOI: 10.1523/jneurosci.4806-10.2011] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
We examined how changes in intensity and interaural time difference (ITD) influenced the coding of low-frequency sounds in the inferior colliculus of male gerbils at both the single neuron and population levels. We found that changes in intensity along the positive slope of the rate-level function (RLF) evoked changes in spectrotemporal filtering that influenced the overall timing of spike events but preserved their precision across trials such that the decoding of single neuron responses was not affected. In contrast, changes in ITD did not trigger changes in spectrotemporal filtering, but did have strong effects on the precision of spike events and, consequently, on decoder performance. However, changes in ITD had opposing effects in the two brain hemispheres and, thus, canceled out at the population level. These results were similar with and without the addition of background noise. We also found that the effects of changes in intensity along the negative slope of the RLF were different from the effects of changes in intensity along the positive slope in that they evoked changes in both spectrotemporal filtering and in the precision of spike events across trials, as well as in decoder performance. These results demonstrate that, at least at moderate intensities, the auditory system employs different strategies at the single neuron and population levels simultaneously to ensure that the coding of sounds is robust to changes in other stimulus features.
Collapse
|
42
|
Abstract
Interaural time differences (ITDs) are the primary cue for the localization of low-frequency sound sources in the azimuthal plane. For decades, it was assumed that the coding of ITDs in the mammalian brain was similar to that in the avian brain, where information is sparsely distributed across individual neurons, but recent studies have suggested otherwise. In this study, we characterized the representation of ITDs in adult male and female gerbils. First, we performed behavioral experiments to determine the acuity with which gerbils can use ITDs to localize sounds. Next, we used different decoders to infer ITDs from the activity of a population of neurons in central nucleus of the inferior colliculus. These results show that ITDs are not represented in a distributed manner, but rather in the summed activity of the entire population. To contrast these results with those from a population where the representation of ITDs is known to be sparsely distributed, we performed the same analysis on activity from the external nucleus of the inferior colliculus of adult male and female barn owls. Together, our results support the idea that, unlike the avian brain, the mammalian brain represents ITDs in the overall activity of a homogenous population of neurons within each hemisphere.
Collapse
|
43
|
Abstract
Neurons in auditory cortex are central to our perception of sounds. However, the underlying neural codes, and the relevance of millisecond-precise spike timing in particular, remain debated. Here, we addressed this issue in the auditory cortex of alert nonhuman primates by quantifying the amount of information carried by precise spike timing about complex sounds presented for extended periods of time (random tone sequences and natural sounds). We investigated the dependence of stimulus information on the temporal precision at which spike times were registered and found that registering spikes at a precision coarser than a few milliseconds significantly reduced the encoded information. This dependence demonstrates that auditory cortex neurons can carry stimulus information at high temporal precision. In addition, we found that the main determinant of finely timed information was rapid modulation of the firing rate, whereas higher-order correlations between spike times contributed negligibly. Although the neural coding precision was high for random tone sequences and natural sounds, the information lost at a precision coarser than a few milliseconds was higher for the stimulus sequence that varied on a faster time scale (random tones), suggesting that the precision of cortical firing depends on the stimulus dynamics. Together, these results provide a neural substrate for recently reported behavioral relevance of precisely timed activity patterns with auditory cortex. In addition, they highlight the importance of millisecond-precise neural coding as general functional principle of auditory processing--from the periphery to cortex.
Collapse
|
44
|
Optimization of population decoding with distance metrics. Neural Netw 2010; 23:728-32. [PMID: 20488662 DOI: 10.1016/j.neunet.2010.04.007] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2009] [Revised: 04/12/2010] [Accepted: 04/27/2010] [Indexed: 11/21/2022]
Abstract
Recent advances in multi-electrode recording and imaging techniques have made it possible to observe the activity of large populations of neurons. However, to take full advantage of these techniques, new methods for the analysis of population responses must be developed. In this paper, we present an algorithm for optimizing population decoding with distance metrics. To demonstrate the utility of this algorithm under experimental conditions, we evaluate its performance in decoding both population spike trains and calcium signals with different correlation structures. Our results demonstrate that the optimized decoder outperforms other simple population decoders and suggest that optimization could serve as a tool for quantifying the potential contribution of individual cells to the population code.
Collapse
|
45
|
Context dependence of spectro-temporal receptive fields with implications for neural coding. Hear Res 2010; 271:123-32. [PMID: 20123121 DOI: 10.1016/j.heares.2010.01.014] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/16/2009] [Revised: 01/25/2010] [Accepted: 01/27/2010] [Indexed: 11/23/2022]
Abstract
The spectro-temporal receptive field (STRF) is frequently used to characterize the linear frequency-time filter properties of the auditory system up to the neuron recorded from. STRFs are extremely stimulus dependent, reflecting the strong non-linearities in the auditory system. Changes in the STRF with stimulus type (tonal, noise-like, vocalizations), sound level and spectro-temporal sound density are reviewed here. Effects on STRF shape of task and attention are also briefly reviewed. Models to account for these changes, potential improvements to STRF analysis, and implications for neural coding are discussed.
Collapse
|
46
|
Rodríguez FA, Read HL, Escabí MA. Spectral and temporal modulation tradeoff in the inferior colliculus. J Neurophysiol 2009; 103:887-903. [PMID: 20018831 DOI: 10.1152/jn.00813.2009] [Citation(s) in RCA: 63] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
The cochlea encodes sounds through frequency-selective channels that exhibit low-pass modulation sensitivity. Unlike the cochlea, neurons in the auditory midbrain are tuned for spectral and temporal modulations found in natural sounds, yet the role of this transformation is not known. We report a distinct tradeoff in modulation sensitivity and tuning that is topographically ordered within the central nucleus of the inferior colliculus (CNIC). Spectrotemporal receptive fields (STRFs) were obtained with 16-channel electrodes inserted orthogonal to the isofrequency lamina. Surprisingly, temporal and spectral characteristics exhibited an opposing relationship along the tonotopic axis. For low best frequencies (BFs), units were selective for fast temporal and broad spectral modulations. A systematic progression was observed toward slower temporal and finer spectral modulation sensitivity at high BF. This tradeoff was strongly reflected in the arrangement of excitation and inhibition and, consequently, in the modulation tuning characteristics. Comparisons with auditory nerve fibers show that these trends oppose the pattern imposed by the peripheral filters. These results suggest that spectrotemporal preferences are reordered within the tonotopic axis of the CNIC. This topographic organization has profound implications for the coding of spectrotemporal features in natural sounds and could underlie a number of perceptual phenomena.
Collapse
|