1
|
Yazaki-Sugiyama Y. Tutor auditory memory for guiding sensorimotor learning in birdsong. Front Neural Circuits 2024; 18:1431119. [PMID: 39011279 PMCID: PMC11246853 DOI: 10.3389/fncir.2024.1431119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2024] [Accepted: 06/13/2024] [Indexed: 07/17/2024] Open
Abstract
Memory-guided motor shaping is necessary for sensorimotor learning. Vocal learning, such as speech development in human babies and song learning in bird juveniles, begins with the formation of an auditory template by hearing adult voices followed by vocally matching to the memorized template using auditory feedback. In zebra finches, the widely used songbird model system, only males develop individually unique stereotyped songs. The production of normal songs relies on auditory experience of tutor's songs (commonly their father's songs) during a critical period in development that consists of orchestrated auditory and sensorimotor phases. "Auditory templates" of tutor songs are thought to form in the brain to guide later vocal learning, while formation of "motor templates" of own song has been suggested to be necessary for the maintenance of stereotyped adult songs. Where these templates are formed in the brain and how they interact with other brain areas to guide song learning, presumably with template-matching error correction, remains to be clarified. Here, we review and discuss studies on auditory and motor templates in the avian brain. We suggest that distinct auditory and motor template systems exist that switch their functions during development.
Collapse
Affiliation(s)
- Yoko Yazaki-Sugiyama
- Neuronal Mechanism for Critical Period Unit, OIST Graduate University, Okinawa, Japan
| |
Collapse
|
2
|
Moseley SM, Meliza CD. Cortical Processing of Conspecific Vocalizations in Zebra Finches Depends on the Early Acoustical Environment. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.06.25.600670. [PMID: 38979160 PMCID: PMC11230381 DOI: 10.1101/2024.06.25.600670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/10/2024]
Abstract
Sensory experience during development has lasting effects on perception and neural processing. Exposing animals to artificial stimuli early in life influences the tuning and functional organization of the auditory cortex, but less is known about how the rich acoustical environments experienced by vocal communicators affect the processing of complex vocalizations. Here, we show that in zebra finches ( Taeniopygia guttata ), a colonial-breeding songbird species, exposure to a naturalistic social-acoustical environment during development has a profound impact on cortical-level auditory responses to conspecific song. Compared to birds raised by pairs in acoustic isolation, birds raised in a breeding colony had higher average firing rates, selectivity, and discriminability, especially in the narrow-spiking, putatively inhibitory neurons of a higher-order auditory area, the caudomedial nidopallium (NCM). Neurons in colony-reared birds were also less correlated in their tuning and more efficient at encoding the spectrotemporal structure of conspecific song. These results suggest that the auditory cortex adapts to noisy, complex acoustical environments by strengthening inhibitory circuitry, functionally decoupling excitatory neurons while maintaining overall excitatory-inhibitory balance.
Collapse
|
3
|
Macedo-Lima M, Fernández-Vargas M, Remage-Healey L. Social reinforcement guides operant behaviour and auditory learning in a songbird. Anim Behav 2024; 210:127-137. [PMID: 38505105 PMCID: PMC10947183 DOI: 10.1016/j.anbehav.2024.01.026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/21/2024]
Abstract
Motivation to seek social interactions is inherent to all social species. For instance, even with risk of disease transmission in a recent pandemic, humans sought out frequent in-person social interactions. In other social animals, socialization can be prioritized even over water or food consumption. Zebra finches, Taeniopygia guttata, are highly gregarious songbirds widely used in behavioural and physiological research. Songbirds, like humans, are vocal learners during development, which rely on intense auditory learning. Aside from supporting song learning, auditory learning further supports individual identification, mate choice and outcome associations in songbirds. To study auditory learning in a laboratory setting, studies often employ operant paradigms with food restriction and reinforcement and require complete social isolation, which can result in stress and other unintended physiological consequences for social species. Thus, in this work, we designed an operant behavioural method leveraging the sociality of zebra finches for goal-directed behaviours. Our approach relies on visual social reinforcement, without depriving the animals of food or social contact. Using this task, we found that visual social reinforcement was a strong motivational drive for operant behaviour. Motivation was sensitive to familiarity towards the stimulus animal and higher when engaging with a familiar versus a novel individual. We further show that this tool can be used to assess auditory discrimination learning using either songs or synthetic pure tones as stimuli. As birds gained experience in the task, they developed a strategy to maximize reward acquisition in spite of receiving more punishment, i.e. liberal response bias. Our operant paradigm provides an alternative to tasks using food reinforcement and could be applied to a variety of highly social species, such as rodents and nonhuman primates.
Collapse
Affiliation(s)
- Matheus Macedo-Lima
- Matheus Macedo-Lima is now at the Department of Biology, University of Maryland, College Park, MD, U.S.A
| | - Marcela Fernández-Vargas
- Marcela Fernández-Vargas is now at the Department of Psychology, Neuroscience Program, Colorado College, Colorado Springs, CO, U.S.A
| | - Luke Remage-Healey
- Corresponding author. (L. Remage-Healey)., @HealeyLab, Neuroscience and Behavior Program, Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst MA, U.S.A.
| |
Collapse
|
4
|
Fernández-Vargas M, Macedo-Lima M, Remage-Healey L. Acute Aromatase Inhibition Impairs Neural and Behavioral Auditory Scene Analysis in Zebra Finches. eNeuro 2024; 11:ENEURO.0423-23.2024. [PMID: 38467426 PMCID: PMC10960633 DOI: 10.1523/eneuro.0423-23.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 12/31/2023] [Accepted: 01/04/2024] [Indexed: 03/13/2024] Open
Abstract
Auditory perception can be significantly disrupted by noise. To discriminate sounds from noise, auditory scene analysis (ASA) extracts the functionally relevant sounds from acoustic input. The zebra finch communicates in noisy environments. Neurons in their secondary auditory pallial cortex (caudomedial nidopallium, NCM) can encode song from background chorus, or scenes, and this capacity may aid behavioral ASA. Furthermore, song processing is modulated by the rapid synthesis of neuroestrogens when hearing conspecific song. To examine whether neuroestrogens support neural and behavioral ASA in both sexes, we retrodialyzed fadrozole (aromatase inhibitor, FAD) and recorded in vivo awake extracellular NCM responses to songs and scenes. We found that FAD affected neural encoding of songs by decreasing responsiveness and timing reliability in inhibitory (narrow-spiking), but not in excitatory (broad-spiking) neurons. Congruently, FAD decreased neural encoding of songs in scenes for both cell types, particularly in females. Behaviorally, we trained birds using operant conditioning and tested their ability to detect songs in scenes after administering FAD orally or injected bilaterally into NCM. Oral FAD increased response bias and decreased correct rejections in females, but not in males. FAD in NCM did not affect performance. Thus, FAD in the NCM impaired neuronal ASA but that did not lead to behavioral disruption suggesting the existence of resilience or compensatory responses. Moreover, impaired performance after systemic FAD suggests involvement of other aromatase-rich networks outside the auditory pathway in ASA. This work highlights how transient estrogen synthesis disruption can modulate higher-order processing in an animal model of vocal communication.
Collapse
Affiliation(s)
- Marcela Fernández-Vargas
- Neuroscience and Behavior Program, Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Matheus Macedo-Lima
- Neuroscience and Behavior Program, Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Luke Remage-Healey
- Neuroscience and Behavior Program, Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| |
Collapse
|
5
|
Alamatsaz N, Rosen MJ, Ihlefeld A. Increased reliance on temporal coding when target sound is softer than the background. Sci Rep 2024; 14:4457. [PMID: 38396044 PMCID: PMC10891139 DOI: 10.1038/s41598-024-54865-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 02/17/2024] [Indexed: 02/25/2024] Open
Abstract
Everyday environments often contain multiple concurrent sound sources that fluctuate over time. Normally hearing listeners can benefit from high signal-to-noise ratios (SNRs) in energetic dips of temporally fluctuating background sound, a phenomenon called dip-listening. Specialized mechanisms of dip-listening exist across the entire auditory pathway. Both the instantaneous fluctuating and the long-term overall SNR shape dip-listening. An unresolved issue regarding cortical mechanisms of dip-listening is how target perception remains invariant to overall SNR, specifically, across different tone levels with an ongoing fluctuating masker. Equivalent target detection over both positive and negative overall SNRs (SNR invariance) is reliably achieved in highly-trained listeners. Dip-listening is correlated with the ability to resolve temporal fine structure, which involves temporally-varying spike patterns. Thus the current work tests the hypothesis that at negative SNRs, neuronal readout mechanisms need to increasingly rely on decoding strategies based on temporal spike patterns, as opposed to spike count. Recordings from chronically implanted electrode arrays in core auditory cortex of trained and awake Mongolian gerbils that are engaged in a tone detection task in 10 Hz amplitude-modulated background sound reveal that rate-based decoding is not SNR-invariant, whereas temporal coding is informative at both negative and positive SNRs.
Collapse
Affiliation(s)
- Nima Alamatsaz
- Graduate School of Biomedical Sciences, Rutgers University, Newark, NJ, USA
- Department of Biomedical Engineering, New Jersey Institute of Technology, Newark, NJ, USA
| | - Merri J Rosen
- Northeast Ohio Medical University (NEOMED), Rootstown, OH, USA.
- University Hospitals Hearing Research Center at NEOMED, Rootstown, OH, USA.
- Brain Health Research Institute, Kent State University, Kent, OH, USA.
| | | |
Collapse
|
6
|
Clonan AC, Zhai X, Stevenson IH, Escabí MA. Low-dimensional interference of mid-level sound statistics predicts human speech recognition in natural environmental noise. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.02.13.579526. [PMID: 38405870 PMCID: PMC10888804 DOI: 10.1101/2024.02.13.579526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/27/2024]
Abstract
Recognizing speech in noise, such as in a busy street or restaurant, is an essential listening task where the task difficulty varies across acoustic environments and noise levels. Yet, current cognitive models are unable to account for changing real-world hearing sensitivity. Here, using natural and perturbed background sounds we demonstrate that spectrum and modulations statistics of environmental backgrounds drastically impact human word recognition accuracy and they do so independently of the noise level. These sound statistics can facilitate or hinder recognition - at the same noise level accuracy can range from 0% to 100%, depending on the background. To explain this perceptual variability, we optimized a biologically grounded hierarchical model, consisting of frequency-tuned cochlear filters and subsequent mid-level modulation-tuned filters that account for central auditory tuning. Low-dimensional summary statistics from the mid-level model accurately predict single trial perceptual judgments, accounting for more than 90% of the perceptual variance across backgrounds and noise levels, and substantially outperforming a cochlear model. Furthermore, perceptual transfer functions in the mid-level auditory space identify multi-dimensional natural sound features that impact recognition. Thus speech recognition in natural backgrounds involves interference of multiple summary statistics that are well described by an interpretable, low-dimensional auditory model. Since this framework relates salient natural sound cues to single trial perceptual judgements, it may improve outcomes for auditory prosthetics and clinical measurements of real-world hearing sensitivity.
Collapse
Affiliation(s)
- Alex C. Clonan
- Electrical and Computer Engineering, University of Connecticut, Storrs, CT 06269
- Biomedical Engineering, University of Connecticut, Storrs, CT 06269
- Institute of Brain and Cognitive Sciences, University of Connecticut, Storrs, CT 06269
| | - Xiu Zhai
- Biomedical Engineering, Wentworth Institute of Technology, Boston, MA 02115
| | - Ian H. Stevenson
- Biomedical Engineering, University of Connecticut, Storrs, CT 06269
- Psychological Sciences, University of Connecticut, Storrs, CT 06269
- Institute of Brain and Cognitive Sciences, University of Connecticut, Storrs, CT 06269
| | - Monty A. Escabí
- Electrical and Computer Engineering, University of Connecticut, Storrs, CT 06269
- Biomedical Engineering, University of Connecticut, Storrs, CT 06269
- Psychological Sciences, University of Connecticut, Storrs, CT 06269
- Institute of Brain and Cognitive Sciences, University of Connecticut, Storrs, CT 06269
| |
Collapse
|
7
|
Martin A, Souffi S, Huetz C, Edeline JM. Can Extensive Training Transform a Mouse into a Guinea Pig? An Evaluation Based on the Discriminative Abilities of Inferior Colliculus Neurons. BIOLOGY 2024; 13:92. [PMID: 38392310 PMCID: PMC10886615 DOI: 10.3390/biology13020092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Revised: 01/19/2024] [Accepted: 01/30/2024] [Indexed: 02/24/2024]
Abstract
Humans and animals maintain accurate discrimination between communication sounds in the presence of loud sources of background noise. In previous studies performed in anesthetized guinea pigs, we showed that, in the auditory pathway, the highest discriminative abilities between conspecific vocalizations were found in the inferior colliculus. Here, we trained CBA/J mice in a Go/No-Go task to discriminate between two similar guinea pig whistles, first in quiet conditions, then in two types of noise, a stationary noise and a chorus noise at three SNRs. Control mice were passively exposed to the same number of whistles as trained mice. After three months of extensive training, inferior colliculus (IC) neurons were recorded under anesthesia and the responses were quantified as in our previous studies. In quiet, the mean values of the firing rate, the temporal reliability and mutual information obtained from trained mice were higher than from the exposed mice and the guinea pigs. In stationary and chorus noise, there were only a few differences between the trained mice and the guinea pigs; and the lowest mean values of the parameters were found in the exposed mice. These results suggest that behavioral training can trigger plasticity in IC that allows mice neurons to reach guinea pig-like discrimination abilities.
Collapse
Affiliation(s)
- Alexandra Martin
- Paris-Saclay Institute of Neuroscience (Neuro-PSI, UMR 9197), CNRS & Université Paris-Saclay, 91400 Saclay, France
| | - Samira Souffi
- Paris-Saclay Institute of Neuroscience (Neuro-PSI, UMR 9197), CNRS & Université Paris-Saclay, 91400 Saclay, France
| | - Chloé Huetz
- Paris-Saclay Institute of Neuroscience (Neuro-PSI, UMR 9197), CNRS & Université Paris-Saclay, 91400 Saclay, France
| | - Jean-Marc Edeline
- Paris-Saclay Institute of Neuroscience (Neuro-PSI, UMR 9197), CNRS & Université Paris-Saclay, 91400 Saclay, France
| |
Collapse
|
8
|
Gao J, Chen H, Fang M, Ding N. Original speech and its echo are segregated and separately processed in the human brain. PLoS Biol 2024; 22:e3002498. [PMID: 38358954 PMCID: PMC10868781 DOI: 10.1371/journal.pbio.3002498] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Accepted: 01/15/2024] [Indexed: 02/17/2024] Open
Abstract
Speech recognition crucially relies on slow temporal modulations (<16 Hz) in speech. Recent studies, however, have demonstrated that the long-delay echoes, which are common during online conferencing, can eliminate crucial temporal modulations in speech but do not affect speech intelligibility. Here, we investigated the underlying neural mechanisms. MEG experiments demonstrated that cortical activity can effectively track the temporal modulations eliminated by an echo, which cannot be fully explained by basic neural adaptation mechanisms. Furthermore, cortical responses to echoic speech can be better explained by a model that segregates speech from its echo than by a model that encodes echoic speech as a whole. The speech segregation effect was observed even when attention was diverted but would disappear when segregation cues, i.e., speech fine structure, were removed. These results strongly suggested that, through mechanisms such as stream segregation, the auditory system can build an echo-insensitive representation of speech envelope, which can support reliable speech recognition.
Collapse
Affiliation(s)
- Jiaxin Gao
- Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, China
| | - Honghua Chen
- Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, China
| | - Mingxuan Fang
- Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, China
| | - Nai Ding
- Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, China
- Nanhu Brain-computer Interface Institute, Hangzhou, China
- The State key Lab of Brain-Machine Intelligence; The MOE Frontier Science Center for Brain Science & Brain-machine Integration, Zhejiang University, Hangzhou, China
| |
Collapse
|
9
|
Nieder A. Convergent Circuit Computation for Categorization in the Brains of Primates and Songbirds. Cold Spring Harb Perspect Biol 2023; 15:a041526. [PMID: 38040453 PMCID: PMC10691494 DOI: 10.1101/cshperspect.a041526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2023]
Abstract
Categorization is crucial for behavioral flexibility because it enables animals to group stimuli into meaningful classes that can easily be generalized to new circumstances. A most abstract quantitative category is set size, the number of elements in a set. This review explores how categorical number representations are realized by the operations of excitatory and inhibitory neurons in associative telencephalic microcircuits in primates and songbirds. Despite the independent evolution of the primate prefrontal cortex and the avian nidopallium caudolaterale, the neuronal computations of these associative pallial circuits show surprising correspondence. Comparing cellular functions in distantly related taxa can inform about the evolutionary principles of circuit computations for cognition in distinctly but convergently realized brain structures.
Collapse
Affiliation(s)
- Andreas Nieder
- Animal Physiology Unit, Institute of Neurobiology, University of Tübingen, 72076 Tübingen, Germany
| |
Collapse
|
10
|
Grijseels DM, Prendergast BJ, Gorman JC, Miller CT. The neurobiology of vocal communication in marmosets. Ann N Y Acad Sci 2023; 1528:13-28. [PMID: 37615212 PMCID: PMC10592205 DOI: 10.1111/nyas.15057] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/25/2023]
Abstract
An increasingly popular animal model for studying the neural basis of social behavior, cognition, and communication is the common marmoset (Callithrix jacchus). Interest in this New World primate across neuroscience is now being driven by their proclivity for prosociality across their repertoire, high volubility, and rapid development, as well as their amenability to naturalistic testing paradigms and freely moving neural recording and imaging technologies. The complement of these characteristics set marmosets up to be a powerful model of the primate social brain in the years to come. Here, we focus on vocal communication because it is the area that has both made the most progress and illustrates the prodigious potential of this species. We review the current state of the field with a focus on the various brain areas and networks involved in vocal perception and production, comparing the findings from marmosets to other animals, including humans.
Collapse
Affiliation(s)
- Dori M Grijseels
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
| | - Brendan J Prendergast
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
| | - Julia C Gorman
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
- Neurosciences Graduate Program, University of California, San Diego, La Jolla, California, USA
| | - Cory T Miller
- Cortical Systems and Behavior Laboratory, University of California, San Diego, La Jolla, California, USA
- Neurosciences Graduate Program, University of California, San Diego, La Jolla, California, USA
| |
Collapse
|
11
|
Nocon JC, Witter J, Gritton H, Han X, Houghton C, Sen K. A robust and compact population code for competing sounds in auditory cortex. J Neurophysiol 2023; 130:775-787. [PMID: 37646080 PMCID: PMC10642980 DOI: 10.1152/jn.00148.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Revised: 08/22/2023] [Accepted: 08/24/2023] [Indexed: 09/01/2023] Open
Abstract
Cortical circuits encoding sensory information consist of populations of neurons, yet how information aggregates via pooling individual cells remains poorly understood. Such pooling may be particularly important in noisy settings where single-neuron encoding is degraded. One example is the cocktail party problem, with competing sounds from multiple spatial locations. How populations of neurons in auditory cortex code competing sounds have not been previously investigated. Here, we apply a novel information-theoretic approach to estimate information in populations of neurons in mouse auditory cortex about competing sounds from multiple spatial locations, including both summed population (SP) and labeled line (LL) codes. We find that a small subset of neurons is sufficient to nearly maximize mutual information over different spatial configurations, with the labeled line code outperforming the summed population code and approaching information levels attained in the absence of competing stimuli. Finally, information in the labeled line code increases with spatial separation between target and masker, in correspondence with behavioral results on spatial release from masking in humans and animals. Taken together, our results reveal that a compact population of neurons in auditory cortex provides a robust code for competing sounds from different spatial locations.NEW & NOTEWORTHY Little is known about how populations of neurons within cortical circuits encode sensory stimuli in the presence of competing stimuli at other spatial locations. Here, we investigate this problem in auditory cortex using a recently proposed information-theoretic approach. We find a small subset of neurons nearly maximizes information about target sounds in the presence of competing maskers, approaching information levels for isolated stimuli, and provides a noise-robust code for sounds in a complex auditory scene.
Collapse
Affiliation(s)
- Jian Carlo Nocon
- Neurophotonics Center, Boston University, Boston, Massachusetts, United States
- Center for Systems Neuroscience, Boston University, Boston, Massachusetts, United States
- Hearing Research Center, Boston University, Boston, Massachusetts, United States
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States
| | - Jake Witter
- Department of Computer Science, University of Bristol, Bristol, United Kingdom
| | - Howard Gritton
- Department of Comparative Biosciences, University of Illinois, Urbana, Illinois, United States
- Department of Bioengineering, University of Illinois, Urbana, Illinois, United States
| | - Xue Han
- Neurophotonics Center, Boston University, Boston, Massachusetts, United States
- Center for Systems Neuroscience, Boston University, Boston, Massachusetts, United States
- Hearing Research Center, Boston University, Boston, Massachusetts, United States
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States
| | - Conor Houghton
- Department of Computer Science, University of Bristol, Bristol, United Kingdom
| | - Kamal Sen
- Neurophotonics Center, Boston University, Boston, Massachusetts, United States
- Center for Systems Neuroscience, Boston University, Boston, Massachusetts, United States
- Hearing Research Center, Boston University, Boston, Massachusetts, United States
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts, United States
| |
Collapse
|
12
|
Liu W, Vicario DS. Dynamic encoding of phonetic categories in zebra finch auditory forebrain. Sci Rep 2023; 13:11172. [PMID: 37430030 DOI: 10.1038/s41598-023-37982-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2023] [Accepted: 06/30/2023] [Indexed: 07/12/2023] Open
Abstract
Vocal communication requires the formation of acoustic categories to enable invariant representations of sounds despite superficial variations. Humans form acoustic categories for speech phonemes, enabling the listener to recognize words independent of speakers; animals can also discriminate speech phonemes. We investigated the neural mechanisms of this process using electrophysiological recordings from the zebra finch secondary auditory area, caudomedial nidopallium (NCM), during passive exposure to human speech stimuli consisting of two naturally spoken words produced by multiple speakers. Analysis of neural distance and decoding accuracy showed improvements in neural discrimination between word categories over the course of exposure, and this improved representation transferred to the same words by novel speakers. We conclude that NCM neurons formed generalized representations of word categories independent of speaker-specific variations that became more refined over the course of passive exposure. The discovery of this dynamic encoding process in NCM suggests a general processing mechanism for forming categorical representations of complex acoustic signals that humans share with other animals.
Collapse
Affiliation(s)
- Wanyi Liu
- Department of Psychology, Rutgers, The State University of New Jersey, Piscataway, NJ, 08854, USA.
| | - David S Vicario
- Department of Psychology, Rutgers, The State University of New Jersey, Piscataway, NJ, 08854, USA.
| |
Collapse
|
13
|
DiTullio RW, Parthiban C, Piasini E, Chaudhari P, Balasubramanian V, Cohen YE. Time as a supervisor: temporal regularity and auditory object learning. Front Comput Neurosci 2023; 17:1150300. [PMID: 37216064 PMCID: PMC10192587 DOI: 10.3389/fncom.2023.1150300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Accepted: 03/30/2023] [Indexed: 05/24/2023] Open
Abstract
Sensory systems appear to learn to transform incoming sensory information into perceptual representations, or "objects," that can inform and guide behavior with minimal explicit supervision. Here, we propose that the auditory system can achieve this goal by using time as a supervisor, i.e., by learning features of a stimulus that are temporally regular. We will show that this procedure generates a feature space sufficient to support fundamental computations of auditory perception. In detail, we consider the problem of discriminating between instances of a prototypical class of natural auditory objects, i.e., rhesus macaque vocalizations. We test discrimination in two ethologically relevant tasks: discrimination in a cluttered acoustic background and generalization to discriminate between novel exemplars. We show that an algorithm that learns these temporally regular features affords better or equivalent discrimination and generalization than conventional feature-selection algorithms, i.e., principal component analysis and independent component analysis. Our findings suggest that the slow temporal features of auditory stimuli may be sufficient for parsing auditory scenes and that the auditory brain could utilize these slowly changing temporal features.
Collapse
Affiliation(s)
- Ronald W. DiTullio
- David Rittenhouse Laboratory, Department of Physics and Astronomy, University of Pennsylvania, Philadelphia, PA, United States
- Neuroscience Graduate Group, University of Pennsylvania, Philadelphia, PA, United States
- Computational Neuroscience Initiative, University of Pennsylvania, Philadelphia, PA, United States
| | - Chetan Parthiban
- David Rittenhouse Laboratory, Department of Physics and Astronomy, University of Pennsylvania, Philadelphia, PA, United States
| | - Eugenio Piasini
- Computational Neuroscience Initiative, University of Pennsylvania, Philadelphia, PA, United States
- Scuola Internazionale Superiore di Studi Avanzati (SISSA), Trieste, Italy
| | - Pratik Chaudhari
- Department of Electrical and Systems Engineering, University of Pennsylvania, Philadelphia, PA, United States
| | - Vijay Balasubramanian
- David Rittenhouse Laboratory, Department of Physics and Astronomy, University of Pennsylvania, Philadelphia, PA, United States
- Computational Neuroscience Initiative, University of Pennsylvania, Philadelphia, PA, United States
- Santa Fe Institute, Santa Fe, NM, United States
| | - Yale E. Cohen
- Departments of Otorhinolaryngology, Neuroscience, and Bioengineering, University of Pennsylvania, Philadelphia, PA, United States
| |
Collapse
|
14
|
Spool JA, Lally AP, Remage-Healey L. Top-down, auditory pallial regulation of the social behavior network. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.03.08.531754. [PMID: 36945416 PMCID: PMC10028912 DOI: 10.1101/2023.03.08.531754] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
UNLABELLED Social encounters rely on sensory cues that carry nuanced information to guide social decision-making. While high-level features of social signals are processed in the telencephalic pallium, nuclei controlling social behaviors, called the social behavior network (SBN), reside mainly in the diencephalon. Although it is well known how mammalian olfactory pallium interfaces with the SBN, there is little information for how pallial processing of other sensory modalities can modulate SBN circuits. This is surprising given the importance of complex vocalizations, for example, for social behavior in many vertebrate taxa such as humans and birds. Using gregarious and highly vocal songbirds, female Zebra finches, we asked to what extent auditory pallial circuits provide consequential input to the SBN as it processes social sensory cues. We transiently inactivated auditory pallium of female Zebra finches during song playback and examined song-induced activation in SBN nuclei. Auditory pallial inactivation impaired responses to song specifically within the lateral ventromedial nucleus of the hypothalamus (VMHl), providing the first evidence in vertebrates of a connection between auditory pallium and the SBN. This same treatment elevated feeding behavior, which also correlated with VMHl activation. This suggests that signals from auditory pallium to VMHl can tune the balance between social attention and feeding drive. A descending influence of sensory pallium on hypothalamic circuits could therefore provide a functional connection for the integration of social stimuli with internal state to influence social decision-making. SIGNIFICANCE Sensory cues such as vocalizations contain important social information. These social signals can be substantially nuanced, containing information about vocalizer identity, prior experience, valence, and emotional state. Processing these features of vocalizations necessitates processing the fast, complex sound streams in song or speech, which depends on circuits in pallial cortex. But whether and how this information is then transferred to social circuits in limbic and hypothalamic regions remains a mystery. Here, we identify a top-down influence of the songbird auditory pallium on one specific node of the social behavior network within the hypothalamus. Descending functional connections such as these may be critical for the wide range of vertebrate species that rely on intricate sensory communication signals to guide social decision-making.
Collapse
|
15
|
Robotka H, Thomas L, Yu K, Wood W, Elie JE, Gahr M, Theunissen FE. Sparse ensemble neural code for a complete vocal repertoire. Cell Rep 2023; 42:112034. [PMID: 36696266 PMCID: PMC10363576 DOI: 10.1016/j.celrep.2023.112034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2022] [Revised: 08/08/2022] [Accepted: 01/09/2023] [Indexed: 01/26/2023] Open
Abstract
The categorization of animal vocalizations into distinct behaviorally relevant groups for communication is an essential operation that must be performed by the auditory system. This auditory object recognition is a difficult task that requires selectivity to the group identifying acoustic features and invariance to renditions within each group. We find that small ensembles of auditory neurons in the forebrain of a social songbird can code the bird's entire vocal repertoire (∼10 call types). Ensemble neural discrimination is not, however, correlated with single unit selectivity, but instead with how well the joint single unit tunings to characteristic spectro-temporal modulations span the acoustic subspace optimized for the discrimination of call types. Thus, akin to face recognition in the visual system, call type recognition in the auditory system is based on a sparse code representing a small number of high-level features and not on highly selective grandmother neurons.
Collapse
Affiliation(s)
- H Robotka
- Max Planck Institute for Ornithology, Seewiesen, Germany
| | - L Thomas
- University of California, Berkeley, Helen Wills Neuroscience Institute, Berkeley, CA, USA
| | - K Yu
- University of California, Berkeley, Helen Wills Neuroscience Institute, Berkeley, CA, USA
| | - W Wood
- University of California, Berkeley, Helen Wills Neuroscience Institute, Berkeley, CA, USA
| | - J E Elie
- University of California, Berkeley, Helen Wills Neuroscience Institute, Berkeley, CA, USA
| | - M Gahr
- Max Planck Institute for Ornithology, Seewiesen, Germany
| | - F E Theunissen
- Max Planck Institute for Ornithology, Seewiesen, Germany; University of California, Berkeley, Helen Wills Neuroscience Institute, Berkeley, CA, USA; Department of Psychology and Integrative Biology, University of California, Berkeley, Berkeley, CA, USA.
| |
Collapse
|
16
|
Mischler G, Keshishian M, Bickel S, Mehta AD, Mesgarani N. Deep neural networks effectively model neural adaptation to changing background noise and suggest nonlinear noise filtering methods in auditory cortex. Neuroimage 2023; 266:119819. [PMID: 36529203 PMCID: PMC10510744 DOI: 10.1016/j.neuroimage.2022.119819] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2022] [Revised: 11/28/2022] [Accepted: 12/15/2022] [Indexed: 12/23/2022] Open
Abstract
The human auditory system displays a robust capacity to adapt to sudden changes in background noise, allowing for continuous speech comprehension despite changes in background environments. However, despite comprehensive studies characterizing this ability, the computations that underly this process are not well understood. The first step towards understanding a complex system is to propose a suitable model, but the classical and easily interpreted model for the auditory system, the spectro-temporal receptive field (STRF), cannot match the nonlinear neural dynamics involved in noise adaptation. Here, we utilize a deep neural network (DNN) to model neural adaptation to noise, illustrating its effectiveness at reproducing the complex dynamics at the levels of both individual electrodes and the cortical population. By closely inspecting the model's STRF-like computations over time, we find that the model alters both the gain and shape of its receptive field when adapting to a sudden noise change. We show that the DNN model's gain changes allow it to perform adaptive gain control, while the spectro-temporal change creates noise filtering by altering the inhibitory region of the model's receptive field. Further, we find that models of electrodes in nonprimary auditory cortex also exhibit noise filtering changes in their excitatory regions, suggesting differences in noise filtering mechanisms along the cortical hierarchy. These findings demonstrate the capability of deep neural networks to model complex neural adaptation and offer new hypotheses about the computations the auditory cortex performs to enable noise-robust speech perception in real-world, dynamic environments.
Collapse
Affiliation(s)
- Gavin Mischler
- Mortimer B. Zuckerman Mind Brain Behavior, Columbia University, New York, United States; Department of Electrical Engineering, Columbia University, New York, United States
| | - Menoua Keshishian
- Mortimer B. Zuckerman Mind Brain Behavior, Columbia University, New York, United States; Department of Electrical Engineering, Columbia University, New York, United States
| | - Stephan Bickel
- Hofstra Northwell School of Medicine, Manhasset, New York, United States
| | - Ashesh D Mehta
- Hofstra Northwell School of Medicine, Manhasset, New York, United States
| | - Nima Mesgarani
- Mortimer B. Zuckerman Mind Brain Behavior, Columbia University, New York, United States; Department of Electrical Engineering, Columbia University, New York, United States.
| |
Collapse
|
17
|
A Redundant Cortical Code for Speech Envelope. J Neurosci 2023; 43:93-112. [PMID: 36379706 PMCID: PMC9838705 DOI: 10.1523/jneurosci.1616-21.2022] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2021] [Revised: 08/19/2022] [Accepted: 10/23/2022] [Indexed: 11/17/2022] Open
Abstract
Animal communication sounds exhibit complex temporal structure because of the amplitude fluctuations that comprise the sound envelope. In human speech, envelope modulations drive synchronized activity in auditory cortex (AC), which correlates strongly with comprehension (Giraud and Poeppel, 2012; Peelle and Davis, 2012; Haegens and Zion Golumbic, 2018). Studies of envelope coding in single neurons, performed in nonhuman animals, have focused on periodic amplitude modulation (AM) stimuli and use response metrics that are not easy to juxtapose with data from humans. In this study, we sought to bridge these fields. Specifically, we looked directly at the temporal relationship between stimulus envelope and spiking, and we assessed whether the apparent diversity across neurons' AM responses contributes to the population representation of speech-like sound envelopes. We gathered responses from single neurons to vocoded speech stimuli and compared them to sinusoidal AM responses in auditory cortex (AC) of alert, freely moving Mongolian gerbils of both sexes. While AC neurons displayed heterogeneous tuning to AM rate, their temporal dynamics were stereotyped. Preferred response phases accumulated near the onsets of sinusoidal AM periods for slower rates (<8 Hz), and an over-representation of amplitude edges was apparent in population responses to both sinusoidal AM and vocoded speech envelopes. Crucially, this encoding bias imparted a decoding benefit: a classifier could discriminate vocoded speech stimuli using summed population activity, while higher frequency modulations required a more sophisticated decoder that tracked spiking responses from individual cells. Together, our results imply that the envelope structure relevant to parsing an acoustic stream could be read-out from a distributed, redundant population code.SIGNIFICANCE STATEMENT Animal communication sounds have rich temporal structure and are often produced in extended sequences, including the syllabic structure of human speech. Although the auditory cortex (AC) is known to play a crucial role in representing speech syllables, the contribution of individual neurons remains uncertain. Here, we characterized the representations of both simple, amplitude-modulated sounds and complex, speech-like stimuli within a broad population of cortical neurons, and we found an overrepresentation of amplitude edges. Thus, a phasic, redundant code in auditory cortex can provide a mechanistic explanation for segmenting acoustic streams like human speech.
Collapse
|
18
|
Yeh YT, Rivera M, Woolley SMN. Auditory sensitivity and vocal acoustics in five species of estrildid songbirds. Anim Behav 2023; 195:107-116. [PMID: 36582925 PMCID: PMC9794177 DOI: 10.1016/j.anbehav.2022.11.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
Comparative studies of acoustic communication in clades with diverse signal features provide a powerful framework for testing relationships between perception and behaviour. We measured auditory sensitivity in five species of estrildid songbirds with acoustically distinct songs and tested whether differences aligned with species differences in song frequency content. Species were chosen based on phylogeny and differences in song acoustics. Behavioural audiograms were obtained using operant training and testing. Adult audiograms were compared across species and between sexes within a species. Juvenile and adult audiograms were compared in one species. The audiograms of adults reared by their own species and those reared and tutored by another species were compared in one species. Results showed that audiograms were similar across species and similar to previous reports of songbird auditory sensitivity. Species differed in the highest frequency detected and the frequency of peak sensitivity. While hearing frequency range was not correlated with song frequency bandwidth, the frequency of peak sensitivity was highly corelated with the frequency of peak energy in song. Sensitivity did not differ based on sex, age or tutoring experience. Our findings suggest that adaptations in songbird auditory sensitivity are largely constrained by shared peripheral and central encoding mechanisms, with species-specific perception appearing only at peak sensitivity.
Collapse
Affiliation(s)
- Yow-Tyng Yeh
- Mortimer B. Zuckerman Mind, Brain, and Behavior Institute, Columbia University, New York, NY, U.S.A
- Department of Biological Sciences, Columbia University, New York, NY, U.S.A
| | - Moises Rivera
- Mortimer B. Zuckerman Mind, Brain, and Behavior Institute, Columbia University, New York, NY, U.S.A
- Department of Psychology, Columbia University, New York, NY, U.S.A
- Department of Psychology, Hunter College and the Graduate Center, City University of New York, New York, NY, U.S.A
| | - Sarah M. N. Woolley
- Mortimer B. Zuckerman Mind, Brain, and Behavior Institute, Columbia University, New York, NY, U.S.A
- Department of Psychology, Columbia University, New York, NY, U.S.A
| |
Collapse
|
19
|
Andreeva IG, Ogorodnikova EA. Auditory Adaptation to Speech Signal Characteristics. J EVOL BIOCHEM PHYS+ 2022. [DOI: 10.1134/s0022093022050027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
20
|
Liu XP, Wang X. Distinct neuronal types contribute to hybrid temporal encoding strategies in primate auditory cortex. PLoS Biol 2022; 20:e3001642. [PMID: 35613218 PMCID: PMC9132345 DOI: 10.1371/journal.pbio.3001642] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2021] [Accepted: 04/22/2022] [Indexed: 11/18/2022] Open
Abstract
Studies of the encoding of sensory stimuli by the brain often consider recorded neurons as a pool of identical units. Here, we report divergence in stimulus-encoding properties between subpopulations of cortical neurons that are classified based on spike timing and waveform features. Neurons in auditory cortex of the awake marmoset (Callithrix jacchus) encode temporal information with either stimulus-synchronized or nonsynchronized responses. When we classified single-unit recordings using either a criteria-based or an unsupervised classification method into regular-spiking, fast-spiking, and bursting units, a subset of intrinsically bursting neurons formed the most highly synchronized group, with strong phase-locking to sinusoidal amplitude modulation (SAM) that extended well above 20 Hz. In contrast with other unit types, these bursting neurons fired primarily on the rising phase of SAM or the onset of unmodulated stimuli, and preferred rapid stimulus onset rates. Such differentiating behavior has been previously reported in bursting neuron models and may reflect specializations for detection of acoustic edges. These units responded to natural stimuli (vocalizations) with brief and precise spiking at particular time points that could be decoded with high temporal stringency. Regular-spiking units better reflected the shape of slow modulations and responded more selectively to vocalizations with overall firing rate increases. Population decoding using time-binned neural activity found that decoding behavior differed substantially between regular-spiking and bursting units. A relatively small pool of bursting units was sufficient to identify the stimulus with high accuracy in a manner that relied on the temporal pattern of responses. These unit type differences may contribute to parallel and complementary neural codes. Neurons in auditory cortex show highly diverse responses to sounds. This study suggests that neuronal type inferred from baseline firing properties accounts for much of this diversity, with a subpopulation of bursting units being specialized for precise temporal encoding.
Collapse
Affiliation(s)
- Xiao-Ping Liu
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, Maryland, United States of America
- * E-mail: (X-PL); (XW)
| | - Xiaoqin Wang
- Laboratory of Auditory Neurophysiology, Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, Maryland, United States of America
- * E-mail: (X-PL); (XW)
| |
Collapse
|
21
|
Spool JA, Bergan JF, Remage-Healey L. A neural circuit perspective on brain aromatase. Front Neuroendocrinol 2022; 65:100973. [PMID: 34942232 PMCID: PMC9667830 DOI: 10.1016/j.yfrne.2021.100973] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/02/2021] [Revised: 12/14/2021] [Accepted: 12/16/2021] [Indexed: 12/23/2022]
Abstract
This review explores the role of aromatase in the brain as illuminated by a set of conserved network-level connections identified in several vertebrate taxa. Aromatase-expressing neurons are neurochemically heterogeneous but the brain regions in which they are found are highly-conserved across the vertebrate lineage. During development, aromatase neurons have a prominent role in sexual differentiation of the brain and resultant sex differences in behavior and human brain diseases. Drawing on literature primarily from birds and rodents, we delineate brain regions that express aromatase and that are strongly interconnected, and suggest that, in many species, aromatase expression essentially defines the Social Behavior Network. Moreover, in several cases the inputs to and outputs from this core Social Behavior Network also express aromatase. Recent advances in molecular and genetic tools for neuroscience now enable in-depth and taxonomically diverse studies of the function of aromatase at the neural circuit level.
Collapse
Affiliation(s)
- Jeremy A Spool
- Center for Neuroendocrine Studies, Neuroscience and Behavior Graduate Program, University of Massachusetts, Amherst, MA 01003, United States
| | - Joseph F Bergan
- Center for Neuroendocrine Studies, Neuroscience and Behavior Graduate Program, University of Massachusetts, Amherst, MA 01003, United States
| | - Luke Remage-Healey
- Center for Neuroendocrine Studies, Neuroscience and Behavior Graduate Program, University of Massachusetts, Amherst, MA 01003, United States.
| |
Collapse
|
22
|
Cell-type specific pallial circuits shape categorical tuning responses in the crow telencephalon. Commun Biol 2022; 5:269. [PMID: 35338240 PMCID: PMC8956685 DOI: 10.1038/s42003-022-03208-z] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2021] [Accepted: 02/28/2022] [Indexed: 01/26/2023] Open
Abstract
The nidopallium caudolaterale (NCL), an integration centre in the telencephalon of birds, plays a crucial role in representing and maintaining abstract categories and concepts. However, the computational principles allowing pallial microcircuits consisting of excitatory and inhibitory neurons to shape the tuning to abstract categories remain elusive. Here we identified the major pallial cell types, putative excitatory projection cells and inhibitory interneurons, by characterizing the waveforms of action potentials recorded in crows performing a cognitively demanding numerical categorization task. Both cell types showed clear differences in their capacity to encode categorical information. Nearby and functionally coupled putative projection neurons generally exhibited similar tuning, whereas putative interneurons showed mainly opposite tuning. The results favour feedforward mechanisms for the shaping of categorical tuning in microcircuits of the NCL. Our findings help to decipher the workings of pallial microcircuits in birds during complex cognition and to compare them vis-a-vis neocortical processes in mammals. Neural recordings from the caudolateral nidopallium in crows during a numerosity task suggest there are two subsets of projection neurons and inhibitory interneurons involved in complex cognition.
Collapse
|
23
|
Lakunina AA, Menashe N, Jaramillo S. Contributions of Distinct Auditory Cortical Inhibitory Neuron Types to the Detection of Sounds in Background Noise. eNeuro 2022; 9:ENEURO.0264-21.2021. [PMID: 35168950 PMCID: PMC8906447 DOI: 10.1523/eneuro.0264-21.2021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2021] [Revised: 10/17/2021] [Accepted: 12/28/2021] [Indexed: 12/01/2022] Open
Abstract
The ability to separate background noise from relevant acoustic signals is essential for appropriate sound-driven behavior in natural environments. Examples of this separation are apparent in the auditory system, where neural responses to behaviorally relevant stimuli become increasingly noise invariant along the ascending auditory pathway. However, the mechanisms that underlie this reduction in responses to background noise are not well understood. To address this gap in knowledge, we first evaluated the effects of auditory cortical inactivation on mice of both sexes trained to perform a simple auditory signal-in-noise detection task and found that outputs from the auditory cortex are important for the detection of auditory stimuli in noisy environments. Next, we evaluated the contributions of the two most common cortical inhibitory cell types, parvalbumin-expressing (PV+) and somatostatin-expressing (SOM+) interneurons, to the perception of masked auditory stimuli. We found that inactivation of either PV+ or SOM+ cells resulted in a reduction in the ability of mice to determine the presence of auditory stimuli masked by noise. These results indicate that a disruption of auditory cortical network dynamics by either of these two types of inhibitory cells is sufficient to impair the ability to separate acoustic signals from noise.
Collapse
Affiliation(s)
- Anna A Lakunina
- Institute of Neuroscience and Department of Biology, University of Oregon, Eugene, Oregon 97403
| | - Nadav Menashe
- Institute of Neuroscience and Department of Biology, University of Oregon, Eugene, Oregon 97403
| | - Santiago Jaramillo
- Institute of Neuroscience and Department of Biology, University of Oregon, Eugene, Oregon 97403
| |
Collapse
|
24
|
Auerbach BD, Gritton HJ. Hearing in Complex Environments: Auditory Gain Control, Attention, and Hearing Loss. Front Neurosci 2022; 16:799787. [PMID: 35221899 PMCID: PMC8866963 DOI: 10.3389/fnins.2022.799787] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Accepted: 01/18/2022] [Indexed: 12/12/2022] Open
Abstract
Listening in noisy or complex sound environments is difficult for individuals with normal hearing and can be a debilitating impairment for those with hearing loss. Extracting meaningful information from a complex acoustic environment requires the ability to accurately encode specific sound features under highly variable listening conditions and segregate distinct sound streams from multiple overlapping sources. The auditory system employs a variety of mechanisms to achieve this auditory scene analysis. First, neurons across levels of the auditory system exhibit compensatory adaptations to their gain and dynamic range in response to prevailing sound stimulus statistics in the environment. These adaptations allow for robust representations of sound features that are to a large degree invariant to the level of background noise. Second, listeners can selectively attend to a desired sound target in an environment with multiple sound sources. This selective auditory attention is another form of sensory gain control, enhancing the representation of an attended sound source while suppressing responses to unattended sounds. This review will examine both “bottom-up” gain alterations in response to changes in environmental sound statistics as well as “top-down” mechanisms that allow for selective extraction of specific sound features in a complex auditory scene. Finally, we will discuss how hearing loss interacts with these gain control mechanisms, and the adaptive and/or maladaptive perceptual consequences of this plasticity.
Collapse
Affiliation(s)
- Benjamin D. Auerbach
- Department of Molecular and Integrative Physiology, Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL, United States
- Neuroscience Program, University of Illinois at Urbana-Champaign, Urbana, IL, United States
- *Correspondence: Benjamin D. Auerbach,
| | - Howard J. Gritton
- Neuroscience Program, University of Illinois at Urbana-Champaign, Urbana, IL, United States
- Department of Comparative Biosciences, University of Illinois at Urbana-Champaign, Urbana, IL, United States
- Department of Bioengineering, University of Illinois at Urbana-Champaign, Urbana, IL, United States
| |
Collapse
|
25
|
Shilling-Scrivo K, Mittelstadt J, Kanold PO. Altered Response Dynamics and Increased Population Correlation to Tonal Stimuli Embedded in Noise in Aging Auditory Cortex. J Neurosci 2021; 41:9650-9668. [PMID: 34611028 PMCID: PMC8612470 DOI: 10.1523/jneurosci.0839-21.2021] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2021] [Revised: 09/25/2021] [Accepted: 09/29/2021] [Indexed: 11/21/2022] Open
Abstract
Age-related hearing loss (presbycusis) is a chronic health condition that affects one-third of the world population. One hallmark of presbycusis is a difficulty hearing in noisy environments. Presbycusis can be separated into two components: alterations of peripheral mechanotransduction of sound in the cochlea and central alterations of auditory processing areas of the brain. Although the effects of the aging cochlea in hearing loss have been well studied, the role of the aging brain in hearing loss is less well understood. Therefore, to examine how age-related central processing changes affect hearing in noisy environments, we used a mouse model (Thy1-GCaMP6s X CBA) that has excellent peripheral hearing in old age. We used in vivo two-photon Ca2+ imaging to measure the responses of neuronal populations in auditory cortex (ACtx) of adult (2-6 months, nine male, six female, 4180 neurons) and aging mice (15-17 months, six male, three female, 1055 neurons) while listening to tones in noisy backgrounds. We found that ACtx neurons in aging mice showed larger responses to tones and have less suppressed responses consistent with reduced inhibition. Aging neurons also showed less sensitivity to temporal changes. Population analysis showed that neurons in aging mice showed higher pairwise activity correlations and showed a reduced diversity in responses to sound stimuli. Using neural decoding techniques, we show a loss of information in neuronal populations in the aging brain. Thus, aging not only affects the responses of single neurons but also affects how these neurons jointly represent stimuli.SIGNIFICANCE STATEMENT Aging results in hearing deficits particularly under challenging listening conditions. We show that auditory cortex contains distinct subpopulations of excitatory neurons that preferentially encode different stimulus features and that aging selectively reduces certain subpopulations. We also show that aging increases correlated activity between neurons and thereby reduces the response diversity in auditory cortex. The loss of population response diversity leads to a decrease of stimulus information and deficits in sound encoding, especially in noisy backgrounds. Future work determining the identities of circuits affected by aging could provide new targets for therapeutic strategies.
Collapse
Affiliation(s)
- Kelson Shilling-Scrivo
- Department of Anatomy and Neurobiology, University of Maryland School of Medicine, Baltimore, Maryland 21230
| | - Jonah Mittelstadt
- Department of Biology, University of Maryland, College Park, Maryland 20742
| | - Patrick O Kanold
- Department of Biology, University of Maryland, College Park, Maryland 20742
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland 20215
- Kavli Neuroscience Discovery Institute, Johns Hopkins University, Baltimore, MD 21205
| |
Collapse
|
26
|
Distinct timescales for the neuronal encoding of vocal signals in a high-order auditory area. Sci Rep 2021; 11:19672. [PMID: 34608248 PMCID: PMC8490347 DOI: 10.1038/s41598-021-99135-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Accepted: 09/21/2021] [Indexed: 02/08/2023] Open
Abstract
The ability of the auditory system to selectively recognize natural sound categories while maintaining a certain degree of tolerance towards variations within these categories, which may have functional roles, is thought to be crucial for vocal communication. To date, it is still largely unknown how the balance between tolerance and sensitivity to variations in acoustic signals is coded at a neuronal level. Here, we investigate whether neurons in a high-order auditory area in zebra finches, a songbird species, are sensitive to natural variations in vocal signals by recording their responses to repeated exposures to identical and variant sound sequences. We used the songs of male birds which tend to be highly repetitive with only subtle variations between renditions. When playing these songs to both anesthetized and awake birds, we found that variations between songs did not affect the neuron firing rate but the temporal reliability of responses. This suggests that auditory processing operates on a range of distinct timescales, namely a short one to detect variations in vocal signals, and longer ones that allow the birds to tolerate variations in vocal signal structure and to encode the global context.
Collapse
|
27
|
Souffi S, Nodal FR, Bajo VM, Edeline JM. When and How Does the Auditory Cortex Influence Subcortical Auditory Structures? New Insights About the Roles of Descending Cortical Projections. Front Neurosci 2021; 15:690223. [PMID: 34413722 PMCID: PMC8369261 DOI: 10.3389/fnins.2021.690223] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Accepted: 06/14/2021] [Indexed: 12/28/2022] Open
Abstract
For decades, the corticofugal descending projections have been anatomically well described but their functional role remains a puzzling question. In this review, we will first describe the contributions of neuronal networks in representing communication sounds in various types of degraded acoustic conditions from the cochlear nucleus to the primary and secondary auditory cortex. In such situations, the discrimination abilities of collicular and thalamic neurons are clearly better than those of cortical neurons although the latter remain very little affected by degraded acoustic conditions. Second, we will report the functional effects resulting from activating or inactivating corticofugal projections on functional properties of subcortical neurons. In general, modest effects have been observed in anesthetized and in awake, passively listening, animals. In contrast, in behavioral tasks including challenging conditions, behavioral performance was severely reduced by removing or transiently silencing the corticofugal descending projections. This suggests that the discriminative abilities of subcortical neurons may be sufficient in many acoustic situations. It is only in particularly challenging situations, either due to the task difficulties and/or to the degraded acoustic conditions that the corticofugal descending connections bring additional abilities. Here, we propose that it is both the top-down influences from the prefrontal cortex, and those from the neuromodulatory systems, which allow the cortical descending projections to impact behavioral performance in reshaping the functional circuitry of subcortical structures. We aim at proposing potential scenarios to explain how, and under which circumstances, these projections impact on subcortical processing and on behavioral responses.
Collapse
Affiliation(s)
- Samira Souffi
- Department of Integrative and Computational Neurosciences, Paris-Saclay Institute of Neuroscience (NeuroPSI), UMR CNRS 9197, Paris-Saclay University, Orsay, France
| | - Fernando R Nodal
- Department of Physiology, Anatomy and Genetics, Medical Sciences Division, University of Oxford, Oxford, United Kingdom
| | - Victoria M Bajo
- Department of Physiology, Anatomy and Genetics, Medical Sciences Division, University of Oxford, Oxford, United Kingdom
| | - Jean-Marc Edeline
- Department of Integrative and Computational Neurosciences, Paris-Saclay Institute of Neuroscience (NeuroPSI), UMR CNRS 9197, Paris-Saclay University, Orsay, France
| |
Collapse
|
28
|
Homma NY, Bajo VM. Lemniscal Corticothalamic Feedback in Auditory Scene Analysis. Front Neurosci 2021; 15:723893. [PMID: 34489635 PMCID: PMC8417129 DOI: 10.3389/fnins.2021.723893] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2021] [Accepted: 07/30/2021] [Indexed: 12/15/2022] Open
Abstract
Sound information is transmitted from the ear to central auditory stations of the brain via several nuclei. In addition to these ascending pathways there exist descending projections that can influence the information processing at each of these nuclei. A major descending pathway in the auditory system is the feedback projection from layer VI of the primary auditory cortex (A1) to the ventral division of medial geniculate body (MGBv) in the thalamus. The corticothalamic axons have small glutamatergic terminals that can modulate thalamic processing and thalamocortical information transmission. Corticothalamic neurons also provide input to GABAergic neurons of the thalamic reticular nucleus (TRN) that receives collaterals from the ascending thalamic axons. The balance of corticothalamic and TRN inputs has been shown to refine frequency tuning, firing patterns, and gating of MGBv neurons. Therefore, the thalamus is not merely a relay stage in the chain of auditory nuclei but does participate in complex aspects of sound processing that include top-down modulations. In this review, we aim (i) to examine how lemniscal corticothalamic feedback modulates responses in MGBv neurons, and (ii) to explore how the feedback contributes to auditory scene analysis, particularly on frequency and harmonic perception. Finally, we will discuss potential implications of the role of corticothalamic feedback in music and speech perception, where precise spectral and temporal processing is essential.
Collapse
Affiliation(s)
- Natsumi Y. Homma
- Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA, United States
- Coleman Memorial Laboratory, Department of Otolaryngology – Head and Neck Surgery, University of California, San Francisco, San Francisco, CA, United States
| | - Victoria M. Bajo
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
29
|
Macedo-Lima M, Boyd HM, Remage-Healey L. Dopamine D1 Receptor Activation Drives Plasticity in the Songbird Auditory Pallium. J Neurosci 2021; 41:6050-6069. [PMID: 34083251 PMCID: PMC8276744 DOI: 10.1523/jneurosci.2823-20.2021] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2020] [Revised: 03/09/2021] [Accepted: 04/21/2021] [Indexed: 12/24/2022] Open
Abstract
Vocal learning species must form and extensively hone associations between sounds and social contingencies. In songbirds, dopamine signaling guides song motor production, variability, and motivation, but it is unclear how dopamine regulates fundamental auditory associations for learning new sounds. We hypothesized that dopamine regulates learning in the auditory pallium, in part by interacting with local neuroestradiol signaling. Here, we show that zebra finch auditory neurons frequently coexpress D1 receptor (D1R) protein, neuroestradiol-synthase, GABA, and parvalbumin (PV). Auditory classical conditioning increased neuroplasticity gene induction in D1R-positive neurons. In vitro, D1R pharmacological activation reduced the amplitude of GABAergic and glutamatergic currents and increased the latter's frequency. In vivo, D1R activation reduced the firing of putative interneurons, increased the firing of putative excitatory neurons, and made both neuronal types unable to adapt to novel stimuli. Together, these findings support the hypothesis that dopamine acting via D1Rs modulates auditory association in the songbird sensory pallium.SIGNIFICANCE STATEMENT Our key finding is that auditory forebrain D1 receptors (D1Rs) modulate auditory plasticity, in support of the hypothesis that dopamine modulates the formation of associations between sounds and outcomes. Recent work in songbirds has identified roles for dopamine in driving reinforcement learning and motor variability in song production. This leaves open whether dopamine shapes the initial events that are critical for learning vocalizations, e.g., auditory learning. Our study begins to address this question in the songbird caudomedial nidopallium (NCM), an analog of the mammalian secondary auditory cortex. Our findings indicate that dopamine receptors are important modulators of excitatory/inhibitory balance and sound association learning mechanisms in the NCM, a system that could be a fundamental feature of vertebrate ascending auditory pathways.
Collapse
Affiliation(s)
- Matheus Macedo-Lima
- Neuroscience and Behavior Program
- Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
- CAPES Foundation, Ministry of Education of Brazil, Brasília, DF 70040-020, Brazil
| | - Hannah M Boyd
- Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Luke Remage-Healey
- Neuroscience and Behavior Program
- Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| |
Collapse
|
30
|
Romero S, Hight AE, Clayton KK, Resnik J, Williamson RS, Hancock KE, Polley DB. Cellular and Widefield Imaging of Sound Frequency Organization in Primary and Higher Order Fields of the Mouse Auditory Cortex. Cereb Cortex 2021; 30:1603-1622. [PMID: 31667491 DOI: 10.1093/cercor/bhz190] [Citation(s) in RCA: 54] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
The mouse auditory cortex (ACtx) contains two core fields-primary auditory cortex (A1) and anterior auditory field (AAF)-arranged in a mirror reversal tonotopic gradient. The best frequency (BF) organization and naming scheme for additional higher order fields remain a matter of debate, as does the correspondence between smoothly varying global tonotopy and heterogeneity in local cellular tuning. Here, we performed chronic widefield and two-photon calcium imaging from the ACtx of awake Thy1-GCaMP6s reporter mice. Data-driven parcellation of widefield maps identified five fields, including a previously unidentified area at the ventral posterior extreme of the ACtx (VPAF) and a tonotopically organized suprarhinal auditory field (SRAF) that extended laterally as far as ectorhinal cortex. Widefield maps were stable over time, where single pixel BFs fluctuated by less than 0.5 octaves throughout a 1-month imaging period. After accounting for neuropil signal and frequency tuning strength, BF organization in neighboring layer 2/3 neurons was intermediate to the heterogeneous salt and pepper organization and the highly precise local organization that have each been described in prior studies. Multiscale imaging data suggest there is no ultrasonic field or secondary auditory cortex in the mouse. Instead, VPAF and a dorsal posterior (DP) field emerged as the strongest candidates for higher order auditory areas.
Collapse
Affiliation(s)
- Sandra Romero
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA
| | - Ariel E Hight
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA
| | - Kameron K Clayton
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA
| | - Jennifer Resnik
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA.,Department of Otolaryngology, Harvard Medical School, Boston, MA 02114, USA
| | - Ross S Williamson
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA.,Department of Otolaryngology, Harvard Medical School, Boston, MA 02114, USA
| | - Kenneth E Hancock
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA.,Department of Otolaryngology, Harvard Medical School, Boston, MA 02114, USA
| | - Daniel B Polley
- Eaton-Peabody Laboratories, Massachusetts Eye and Ear Infirmary, Boston, MA 02114, USA.,Department of Otolaryngology, Harvard Medical School, Boston, MA 02114, USA
| |
Collapse
|
31
|
Spool JA, Macedo-Lima M, Scarpa G, Morohashi Y, Yazaki-Sugiyama Y, Remage-Healey L. Genetically identified neurons in avian auditory pallium mirror core principles of their mammalian counterparts. Curr Biol 2021; 31:2831-2843.e6. [PMID: 33989528 DOI: 10.1016/j.cub.2021.04.039] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2020] [Revised: 02/12/2021] [Accepted: 04/15/2021] [Indexed: 12/21/2022]
Abstract
In vertebrates, advanced cognitive abilities are typically associated with the telencephalic pallium. In mammals, the pallium is a layered mixture of excitatory and inhibitory neuronal populations with distinct molecular, physiological, and network phenotypes. This cortical architecture is proposed to support efficient, high-level information processing. Comparative perspectives across vertebrates provide a lens to understand the common features of pallium that are important for advanced cognition. Studies in songbirds have established strikingly parallel features of neuronal types between mammalian and avian pallium. However, lack of genetic access to defined pallial cell types in non-mammalian vertebrates has hindered progress in resolving connections between molecular and physiological phenotypes. A definitive mapping of the physiology of pallial cells onto their molecular identities in birds is critical for understanding how synaptic and computational properties depend on underlying molecular phenotypes. Using viral tools to target excitatory versus inhibitory neurons in the zebra finch auditory association pallium (calmodulin-dependent kinase alpha [CaMKIIα] and glutamate decarboxylase 1 [GAD1] promoters, respectively), we systematically tested predictions derived from mammalian pallium. We identified two genetically distinct neuronal populations that exhibit profound physiological and computational similarities with mammalian excitatory and inhibitory pallial cells, definitively aligning putative cell types in avian caudal nidopallium with these molecular identities. Specifically, genetically identified CaMKIIα and GAD1 cell types in avian auditory association pallium exhibit distinct intrinsic physiological parameters, distinct auditory coding principles, and inhibitory-dependent pallial synchrony, gamma oscillations, and local suppression. The retention, or convergence, of these molecular and physiological features in both birds and mammals clarifies the characteristics of pallial circuits for advanced cognitive abilities.
Collapse
Affiliation(s)
- Jeremy A Spool
- Neuroscience and Behavior, Center for Neuroendocrine Studies, University of Massachusetts, Amherst, MA 01003, USA
| | - Matheus Macedo-Lima
- Neuroscience and Behavior, Center for Neuroendocrine Studies, University of Massachusetts, Amherst, MA 01003, USA; CAPES Foundation, Ministry of Education of Brazil, Brasília 70040-020, Brazil
| | - Garrett Scarpa
- Neuroscience and Behavior, Center for Neuroendocrine Studies, University of Massachusetts, Amherst, MA 01003, USA
| | - Yuichi Morohashi
- Okinawa Institute of Science and Technology (OIST) Graduate University, Okinawa, Japan
| | - Yoko Yazaki-Sugiyama
- Okinawa Institute of Science and Technology (OIST) Graduate University, Okinawa, Japan
| | - Luke Remage-Healey
- Neuroscience and Behavior, Center for Neuroendocrine Studies, University of Massachusetts, Amherst, MA 01003, USA.
| |
Collapse
|
32
|
Boos M, Lücke J, Rieger JW. Generalizable dimensions of human cortical auditory processing of speech in natural soundscapes: A data-driven ultra high field fMRI approach. Neuroimage 2021; 237:118106. [PMID: 33991696 DOI: 10.1016/j.neuroimage.2021.118106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Accepted: 04/25/2021] [Indexed: 11/27/2022] Open
Abstract
Speech comprehension in natural soundscapes rests on the ability of the auditory system to extract speech information from a complex acoustic signal with overlapping contributions from many sound sources. Here we reveal the canonical processing of speech in natural soundscapes on multiple scales by using data-driven modeling approaches to characterize sounds to analyze ultra high field fMRI recorded while participants listened to the audio soundtrack of a movie. We show that at the functional level the neuronal processing of speech in natural soundscapes can be surprisingly low dimensional in the human cortex, highlighting the functional efficiency of the auditory system for a seemingly complex task. Particularly, we find that a model comprising three functional dimensions of auditory processing in the temporal lobes is shared across participants' fMRI activity. We further demonstrate that the three functional dimensions are implemented in anatomically overlapping networks that process different aspects of speech in natural soundscapes. One is most sensitive to complex auditory features present in speech, another to complex auditory features and fast temporal modulations, that are not specific to speech, and one codes mainly sound level. These results were derived with few a-priori assumptions and provide a detailed and computationally reproducible account of the cortical activity in the temporal lobe elicited by the processing of speech in natural soundscapes.
Collapse
Affiliation(s)
- Moritz Boos
- Applied Neurocognitive Psychology Lab, University of Oldenburg, Oldenburg, Germany; Cluster of Excellence "Hearing4all", University of Oldenburg, Oldenburg, Germany.
| | - Jörg Lücke
- Machine Learning Division, University of Oldenburg, Oldenburg, Germany; Cluster of Excellence "Hearing4all", University of Oldenburg, Oldenburg, Germany
| | - Jochem W Rieger
- Applied Neurocognitive Psychology Lab, University of Oldenburg, Oldenburg, Germany; Cluster of Excellence "Hearing4all", University of Oldenburg, Oldenburg, Germany
| |
Collapse
|
33
|
Yao JD, Sanes DH. Temporal Encoding is Required for Categorization, But Not Discrimination. Cereb Cortex 2021; 31:2886-2897. [PMID: 33429423 DOI: 10.1093/cercor/bhaa396] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2020] [Revised: 10/26/2020] [Accepted: 11/03/2020] [Indexed: 11/14/2022] Open
Abstract
Core auditory cortex (AC) neurons encode slow fluctuations of acoustic stimuli with temporally patterned activity. However, whether temporal encoding is necessary to explain auditory perceptual skills remains uncertain. Here, we recorded from gerbil AC neurons while they discriminated between a 4-Hz amplitude modulation (AM) broadband noise and AM rates >4 Hz. We found a proportion of neurons possessed neural thresholds based on spike pattern or spike count that were better than the recorded session's behavioral threshold, suggesting that spike count could provide sufficient information for this perceptual task. A population decoder that relied on temporal information outperformed a decoder that relied on spike count alone, but the spike count decoder still remained sufficient to explain average behavioral performance. This leaves open the possibility that more demanding perceptual judgments require temporal information. Thus, we asked whether accurate classification of different AM rates between 4 and 12 Hz required the information contained in AC temporal discharge patterns. Indeed, accurate classification of these AM stimuli depended on the inclusion of temporal information rather than spike count alone. Overall, our results compare two different representations of time-varying acoustic features that can be accessed by downstream circuits required for perceptual judgments.
Collapse
Affiliation(s)
- Justin D Yao
- Center for Neural Science, New York University, New York, NY 10003, USA
| | - Dan H Sanes
- Center for Neural Science, New York University, New York, NY 10003, USA.,Department of Psychology, New York University, New York, NY 10003, USA.,Department of Biology, New York University, New York, NY 10003, USA.,Neuroscience Institute, NYU Langone Medical Center, New York University, New York, NY 10016, USA
| |
Collapse
|
34
|
Homma NY, Hullett PW, Atencio CA, Schreiner CE. Auditory Cortical Plasticity Dependent on Environmental Noise Statistics. Cell Rep 2021; 30:4445-4458.e5. [PMID: 32234479 PMCID: PMC7326484 DOI: 10.1016/j.celrep.2020.03.014] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2019] [Revised: 08/07/2019] [Accepted: 03/05/2020] [Indexed: 01/14/2023] Open
Abstract
During critical periods, neural circuits develop to form receptive fields that adapt to the sensory environment and enable optimal performance of relevant tasks. We hypothesized that early exposure to background noise can improve signal-in-noise processing, and the resulting receptive field plasticity in the primary auditory cortex can reveal functional principles guiding that important task. We raised rat pups in different spectro-temporal noise statistics during their auditory critical period. As adults, they showed enhanced behavioral performance in detecting vocalizations in noise. Concomitantly, encoding of vocalizations in noise in the primary auditory cortex improves with noise-rearing. Significantly, spectro-temporal modulation plasticity shifts cortical preferences away from the exposed noise statistics, thus reducing noise interference with the foreground sound representation. Auditory cortical plasticity shapes receptive field preferences to optimally extract foreground information in noisy environments during noise-rearing. Early noise exposure induces cortical circuits to implement efficient coding in the joint spectral and temporal modulation domain. After rearing rats in moderately loud spectro-temporally modulated background noise, Homma et al. investigated signal-in-noise processing in the primary auditory cortex. Noise-rearing improved vocalization-in-noise performance in both behavioral testing and neural decoding. Cortical plasticity shifted neuronal spectro-temporal modulation preferences away from the exposed noise statistics.
Collapse
Affiliation(s)
- Natsumi Y Homma
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California, San Francisco, San Francisco, CA 94143, USA; Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA 94143, USA
| | - Patrick W Hullett
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California, San Francisco, San Francisco, CA 94143, USA; Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA 94143, USA
| | - Craig A Atencio
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California, San Francisco, San Francisco, CA 94143, USA; Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA 94143, USA
| | - Christoph E Schreiner
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California, San Francisco, San Francisco, CA 94143, USA; Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA 94143, USA.
| |
Collapse
|
35
|
Robustness to Noise in the Auditory System: A Distributed and Predictable Property. eNeuro 2021; 8:ENEURO.0043-21.2021. [PMID: 33632813 PMCID: PMC7986545 DOI: 10.1523/eneuro.0043-21.2021] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2021] [Revised: 02/17/2021] [Accepted: 02/17/2021] [Indexed: 12/30/2022] Open
Abstract
Background noise strongly penalizes auditory perception of speech in humans or vocalizations in animals. Despite this, auditory neurons are still able to detect communications sounds against considerable levels of background noise. We collected neuronal recordings in cochlear nucleus (CN), inferior colliculus (IC), auditory thalamus, and primary and secondary auditory cortex in response to vocalizations presented either against a stationary or a chorus noise in anesthetized guinea pigs at three signal-to-noise ratios (SNRs; −10, 0, and 10 dB). We provide evidence that, at each level of the auditory system, five behaviors in noise exist within a continuum, from neurons with high-fidelity representations of the signal, mostly found in IC and thalamus, to neurons with high-fidelity representations of the noise, mostly found in CN for the stationary noise and in similar proportions in each structure for the chorus noise. The two cortical areas displayed fewer robust responses than the IC and thalamus. Furthermore, between 21% and 72% of the neurons (depending on the structure) switch categories from one background noise to another, even if the initial assignment of these neurons to a category was confirmed by a severe bootstrap procedure. Importantly, supervised learning pointed out that assigning a recording to one of the five categories can be predicted up to a maximum of 70% based on both the response to signal alone and noise alone.
Collapse
|
36
|
Schroeder KM, Remage-Healey L. Adult-like neural representation of species-specific songs in the auditory forebrain of zebra finch nestlings. Dev Neurobiol 2021; 81:123-138. [PMID: 33369121 PMCID: PMC7969438 DOI: 10.1002/dneu.22802] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2020] [Revised: 11/22/2020] [Accepted: 12/21/2020] [Indexed: 12/30/2022]
Abstract
Encoding of conspecific signals during development can reinforce species barriers as well as set the stage for learning and production of species-typical vocalizations. In altricial songbirds, the development of the auditory system is not complete at hatching, so it is unknown the degree to which recently hatched young can process auditory signals like birdsong. We measured in vivo extracellular responses to song stimuli in a zebra finch (Taeniopygia guttata) secondary auditory forebrain region, the caudomedial nidopallium (NCM). We recorded from three age groups between 13 days post-hatch and adult to identify possible shifts in stimulus encoding that occur before the opening of the sensitive period of song motor learning. We did not find differences in putative cell type composition, firing rate, response strength, and selectivity across ages. Across ages narrow-spiking units had higher firing rates, response strength, accuracy, and trial-by-trial reliability along with lower selectivity than broad-spiking units. In addition, we showed that stimulus-specific adaptation, a characteristic of adult NCM, was also present in nestlings and fledglings. These results indicate that most features of secondary auditory processing are already adult-like shortly after hatching. Furthermore, we showed that selectivity for species-specific stimuli is similar across all ages, with the greatest fidelity in temporal coding in response to conspecific song and domesticated Bengalese finch song, and reduced fidelity in response to owl finch song, a more ecologically relevant heterospecific, and white noise. Our study provides the first evidence that the electrophysiological properties of higher-order auditory neurons are already mature in nestling songbirds.
Collapse
Affiliation(s)
- Katie M. Schroeder
- Organismic and Evolutionary Biology, University of Massachusetts Amherst, Amherst, MA, USA
| | - Luke Remage-Healey
- Organismic and Evolutionary Biology, University of Massachusetts Amherst, Amherst, MA, USA
- Center for Neuroendocrine Studies, University of Massachusetts Amherst, Amherst, MA, USA
| |
Collapse
|
37
|
Hosseini M, Rodriguez G, Guo H, Lim HH, Plourde E. The effect of input noises on the activity of auditory neurons using GLM-based metrics. J Neural Eng 2021; 18. [PMID: 33626516 DOI: 10.1088/1741-2552/abe979] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2020] [Accepted: 02/24/2021] [Indexed: 11/11/2022]
Abstract
CONTEXT The auditory system is extremely efficient in extracting auditory information in the presence of background noise. However, people with auditory implants have a hard time understanding speech in noisy conditions. Understanding the mechanisms of perception in noise could lead to better stimulation or preprocessing strategies for such implants. OBJECTIVE The neural mechanisms related to the processing of background noise, especially in the inferior colliculus (IC) where the auditory midbrain implant is located, are still not well understood. We thus wish to investigate if there is a difference in the activity of neurons in the IC when presenting noisy vocalizations with different types of noise (stationary vs. non-stationary), input signal-to-noise ratios (SNR) and signal levels. APPROACH We developed novel metrics based on a generalized linear model (GLM) to investigate the effect of a given input noise on neural activity. We used these metrics to analyze neural data recorded from the IC in ketamine-anesthetized female Hartley guinea pigs while presenting noisy vocalizations. MAIN RESULTS We found that non-stationary noise clearly contributes to the multi-unit neural activity in the IC by causing excitation, regardless of the SNR, input level or vocalization type. However, when presenting white or natural stationary noises, a great diversity of responses was observed for the different conditions, where the multi-unit activity of some sites was affected by the presence of noise and the activity of others was not. SIGNIFICANCE The GLM-based metrics allowed the identification of a clear distinction between the effect of white or natural stationary noises and that of non-stationary noise on the multi-unit activity in the IC. This had not been observed before and indicates that the so-called noise invariance in the IC is dependent on the input noisy conditions. This could suggest different preprocessing or stimulation approaches for auditory midbrain implants depending on the noisy conditions.
Collapse
Affiliation(s)
- Maryam Hosseini
- Electrical engineering, Université de Sherbrooke, 2500 Boulevard de l'Université, Sherbrooke, Quebec, J1K 2R1, CANADA
| | - Gerardo Rodriguez
- Biomedical engineering, University of Minnesota, 312 Church St SE, Minneapolis, Minnesota, 55455, UNITED STATES
| | - Hongsun Guo
- Biomedical engineering, University of Minnesota, 312 Church St SE, Minneapolis, Minnesota, 55455, UNITED STATES
| | - Hubert H Lim
- Department of Biomedical Engineering, University of Minnesota, 7-105 Hasselmo Hall, 312 Church Street SE, Minneapolis, MN 55455, USA, Minneapolis, Minnesota, 55455, UNITED STATES
| | - Eric Plourde
- Electrical engineering, Université de Sherbrooke, 2500 Boulevard de l'Université, Sherbrooke, Quebec, J1K 2R1, CANADA
| |
Collapse
|
38
|
Pupillometry as a reliable metric of auditory detection and discrimination across diverse stimulus paradigms in animal models. Sci Rep 2021; 11:3108. [PMID: 33542266 PMCID: PMC7862232 DOI: 10.1038/s41598-021-82340-y] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2020] [Accepted: 01/08/2021] [Indexed: 12/30/2022] Open
Abstract
Estimates of detection and discrimination thresholds are often used to explore broad perceptual similarities between human subjects and animal models. Pupillometry shows great promise as a non-invasive, easily-deployable method of comparing human and animal thresholds. Using pupillometry, previous studies in animal models have obtained threshold estimates to simple stimuli such as pure tones, but have not explored whether similar pupil responses can be evoked by complex stimuli, what other stimulus contingencies might affect stimulus-evoked pupil responses, and if pupil responses can be modulated by experience or short-term training. In this study, we used an auditory oddball paradigm to estimate detection and discrimination thresholds across a wide range of stimuli in guinea pigs. We demonstrate that pupillometry yields reliable detection and discrimination thresholds across a range of simple (tones) and complex (conspecific vocalizations) stimuli; that pupil responses can be robustly evoked using different stimulus contingencies (low-level acoustic changes, or higher level categorical changes); and that pupil responses are modulated by short-term training. These results lay the foundation for using pupillometry as a reliable method of estimating thresholds in large experimental cohorts, and unveil the full potential of using pupillometry to explore broad similarities between humans and animal models.
Collapse
|
39
|
The second harmonic neurons in auditory midbrain of Hipposideros pratti are more tolerant to background white noise. Hear Res 2020; 400:108142. [PMID: 33310564 DOI: 10.1016/j.heares.2020.108142] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/13/2020] [Revised: 11/29/2020] [Accepted: 12/02/2020] [Indexed: 12/22/2022]
Abstract
Although acoustic communication is inevitably influenced by noise, behaviorally relevant sounds are perceived reliably. The noise-tolerant and -invariant responses of auditory neurons are thought to be the underlying mechanism. So, it is reasonable to speculate that neurons with best frequency tuned to behaviorally relevant sounds will play important role in noise-tolerant perception. Echolocating bats live in groups and emit multiple harmonic signals and analyze the returning echoes to extract information about the target features, making them prone to deal with noise in their natural habitat. The echolocation signal of Hipposideros pratti usually contains 3-4 harmonics (H1H4), the second harmonic has the highest amplitude and is thought to play an essential role during echolocation behavior. Therefore, it is reasonable to propose that neurons tuned to the H2, named the H2 neurons, can be more noise-tolerant to background noise. Taking advantage of bat's stereotypical echolocation signal and single-cell recording, our present study showed that the minimal threshold increases (12.2 dB) of H2 neurons in the auditory midbrain were comparable to increase in bat's call intensity (14.2 dB) observed in 70 dB SPL white noise condition, indicating that the H2 neurons could work as background noise monitor. The H2 neurons had higher minimal thresholds and sharper frequency tuning, which enabled them to be more tolerant to background noise. Furthermore, the H2 neurons had consistent best amplitude spikes and sharper intensity tuning in background white noise condition than in silence. Taken together, these results suggest that the H2 neurons might account for noise-tolerant perception of behaviorally relevant sounds.
Collapse
|
40
|
Chiang CH, Lee J, Wang C, Williams AJ, Lucas TH, Cohen YE, Viventi J. A modular high-density μECoG system on macaque vlPFC for auditory cognitive decoding. J Neural Eng 2020; 17:046008. [PMID: 32498058 DOI: 10.1088/1741-2552/ab9986] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
Abstract
OBJECTIVE A fundamental goal of the auditory system is to parse the auditory environment into distinct perceptual representations. Auditory perception is mediated by the ventral auditory pathway, which includes the ventrolateral prefrontal cortex (vlPFC). Because large-scale recordings of auditory signals are quite rare, the spatiotemporal resolution of the neuronal code that underlies vlPFC's contribution to auditory perception has not been fully elucidated. Therefore, we developed a modular, chronic, high-resolution, multi-electrode array system with long-term viability in order to identify the information that could be decoded from μECoG vlPFC signals. APPROACH We molded three separate μECoG arrays into one and implanted this system in a non-human primate. A custom 3D-printed titanium chamber was mounted on the left hemisphere. The molded 294-contact μECoG array was implanted subdurally over the vlPFC. μECoG activity was recorded while the monkey participated in a 'hearing-in-noise' task in which they reported hearing a 'target' vocalization from a background 'chorus' of vocalizations. We titrated task difficulty by varying the sound level of the target vocalization, relative to the chorus (target-to-chorus ratio, TCr). MAIN RESULTS We decoded the TCr and the monkey's behavioral choices from the μECoG signal. We analyzed decoding accuracy as a function of number of electrodes, spatial resolution, and time from implantation. Over a one-year period, we found significant decoding with individual electrodes that increased significantly as we decoded simultaneously more electrodes. Further, we found that the decoding for behavioral choice was better than the decoding of TCr. Finally, because the decoding accuracy of individual electrodes varied on a day-by-day basis, electrode arrays with high channel counts ensure robust decoding in the long term. SIGNIFICANCE Our results demonstrate the utility of high-resolution and high-channel-count, chronic µECoG recording. We developed a surface electrode array that can be scaled to cover larger cortical areas without increasing the chamber footprint.
Collapse
Affiliation(s)
- Chia-Han Chiang
- Department of Biomedical Engineering, Duke University, Durham, NC, United States of America. These authors contributed equally to this work
| | | | | | | | | | | | | |
Collapse
|
41
|
Spiking network optimized for word recognition in noise predicts auditory system hierarchy. PLoS Comput Biol 2020; 16:e1007558. [PMID: 32559204 PMCID: PMC7329140 DOI: 10.1371/journal.pcbi.1007558] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2019] [Revised: 07/01/2020] [Accepted: 11/22/2019] [Indexed: 11/21/2022] Open
Abstract
The auditory neural code is resilient to acoustic variability and capable of recognizing sounds amongst competing sound sources, yet, the transformations enabling noise robust abilities are largely unknown. We report that a hierarchical spiking neural network (HSNN) optimized to maximize word recognition accuracy in noise and multiple talkers predicts organizational hierarchy of the ascending auditory pathway. Comparisons with data from auditory nerve, midbrain, thalamus and cortex reveals that the optimal HSNN predicts several transformations of the ascending auditory pathway including a sequential loss of temporal resolution and synchronization ability, increasing sparseness, and selectivity. The optimal organizational scheme enhances performance by selectively filtering out noise and fast temporal cues such as voicing periodicity, that are not directly relevant to the word recognition task. An identical network arranged to enable high information transfer fails to predict auditory pathway organization and has substantially poorer performance. Furthermore, conventional single-layer linear and nonlinear receptive field networks that capture the overall feature extraction of the HSNN fail to achieve similar performance. The findings suggest that the auditory pathway hierarchy and its sequential nonlinear feature extraction computations enhance relevant cues while removing non-informative sources of noise, thus enhancing the representation of sounds in noise impoverished conditions. The brain’s ability to recognize sounds in the presence of competing sounds or background noise is essential for everyday hearing tasks. How the brain accomplishes noise resiliency, however, is poorly understood. Using neural recordings from the ascending auditory pathway and an auditory spiking network model trained for sound recognition in noise we explore the computational strategies that enable noise robustness. Our results suggest that the hierarchical feature organization of the ascending auditory pathway and the resulting computations are critical for sound recognition in the presence of noise.
Collapse
|
42
|
Noise-Sensitive But More Precise Subcortical Representations Coexist with Robust Cortical Encoding of Natural Vocalizations. J Neurosci 2020; 40:5228-5246. [PMID: 32444386 DOI: 10.1523/jneurosci.2731-19.2020] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2019] [Revised: 05/08/2020] [Accepted: 05/15/2020] [Indexed: 01/30/2023] Open
Abstract
Humans and animals maintain accurate sound discrimination in the presence of loud sources of background noise. It is commonly assumed that this ability relies on the robustness of auditory cortex responses. However, only a few attempts have been made to characterize neural discrimination of communication sounds masked by noise at each stage of the auditory system and to quantify the noise effects on the neuronal discrimination in terms of alterations in amplitude modulations. Here, we measured neural discrimination between communication sounds masked by a vocalization-shaped stationary noise from multiunit responses recorded in the cochlear nucleus, inferior colliculus, auditory thalamus, and primary and secondary auditory cortex at several signal-to-noise ratios (SNRs) in anesthetized male or female guinea pigs. Masking noise decreased sound discrimination of neuronal populations in each auditory structure, but collicular and thalamic populations showed better performance than cortical populations at each SNR. In contrast, in each auditory structure, discrimination by neuronal populations was slightly decreased when tone-vocoded vocalizations were tested. These results shed new light on the specific contributions of subcortical structures to robust sound encoding, and suggest that the distortion of slow amplitude modulation cues conveyed by communication sounds is one of the factors constraining the neuronal discrimination in subcortical and cortical levels.SIGNIFICANCE STATEMENT Dissecting how auditory neurons discriminate communication sounds in noise is a major goal in auditory neuroscience. Robust sound coding in noise is often viewed as a specific property of cortical networks, although this remains to be demonstrated. Here, we tested the discrimination performance of neuronal populations at five levels of the auditory system in response to conspecific vocalizations masked by noise. In each acoustic condition, subcortical neurons better discriminated target vocalizations than cortical ones and in each structure, the reduction in discrimination performance was related to the reduction in slow amplitude modulation cues.
Collapse
|
43
|
Dong M, Vicario DS. Statistical learning of transition patterns in the songbird auditory forebrain. Sci Rep 2020; 10:7848. [PMID: 32398864 PMCID: PMC7217825 DOI: 10.1038/s41598-020-64671-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2019] [Accepted: 04/10/2020] [Indexed: 12/04/2022] Open
Abstract
Statistical learning of transition patterns between sounds—a striking capability of the auditory system—plays an essential role in animals’ survival (e.g., detect deviant sounds that signal danger). However, the neural mechanisms underlying this capability are still not fully understood. We recorded extracellular multi-unit and single-unit activity in the auditory forebrain of awake male zebra finches while presenting rare repetitions of a single sound in a long sequence of sounds (canary and zebra finch song syllables) patterned in either an alternating or random order at different inter-stimulus intervals (ISI). When preceding stimuli were regularly alternating (alternating condition), a repeated stimulus violated the preceding transition pattern and was a deviant. When preceding stimuli were in random order (control condition), a repeated stimulus did not violate any regularities and was not a deviant. At all ISIs tested (1 s, 3 s, or jittered at 0.8–1.2 s), deviant repetition enhanced neural responses in the alternating condition in a secondary auditory area (caudomedial nidopallium, NCM) but not in the primary auditory area (Field L2); in contrast, repetition suppressed responses in the control condition in both Field L2 and NCM. When stimuli were presented in the classical oddball paradigm at jittered ISI (0.8–1.2 s), neural responses in both NCM and Field L2 were stronger when a stimulus occurred as deviant with low probability than when the same stimulus occurred as standard with high probability. Together, these results demonstrate: (1) classical oddball effect exists even when ISI is jittered and the onset of a stimulus is not fully predictable; (2) neurons in NCM can learn transition patterns between sounds at multiple ISIs and detect violation of these transition patterns; (3) sensitivity to deviant sounds increases from Field L2 to NCM in the songbird auditory forebrain. Further studies using the current paradigms may help us understand the neural substrate of statistical learning and even speech comprehension.
Collapse
Affiliation(s)
- Mingwen Dong
- Department of Psychology, Rutgers, the State University of New Jersey, New Brunswick, NJ, United States.
| | - David S Vicario
- Department of Psychology, Rutgers, the State University of New Jersey, New Brunswick, NJ, United States
| |
Collapse
|
44
|
Experience-Dependent Coding of Time-Dependent Frequency Trajectories by Off Responses in Secondary Auditory Cortex. J Neurosci 2020; 40:4469-4482. [PMID: 32327533 PMCID: PMC7275866 DOI: 10.1523/jneurosci.2665-19.2020] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2019] [Revised: 04/02/2020] [Accepted: 04/07/2020] [Indexed: 11/21/2022] Open
Abstract
Time-dependent frequency trajectories are an inherent feature of many behaviorally relevant sounds, such as species-specific vocalizations. Dynamic frequency trajectories, even in short sounds, often convey meaningful information, which may be used to differentiate sound categories. However, it is not clear what and where neural responses in the auditory cortical pathway are critical for conveying information about behaviorally relevant frequency trajectories, and how these responses change with experience. Here, we uncover tuning to subtle variations in frequency trajectories in auditory cortex of female mice. We found that auditory cortical responses could be modulated by variations in a pure tone trajectory as small as 1/24th of an octave, comparable to what has been reported in primates. In particular, late spiking after the end of a sound stimulus was more often sensitive to the sound's subtle frequency variation compared with spiking during the sound. Such “Off” responses in the adult A2, but not those in core auditory cortex, were plastic in a way that may enhance the representation of a newly acquired, behaviorally relevant sound category. We illustrate this with the maternal mouse paradigm for natural vocalization learning. By using an ethologically inspired paradigm to drive auditory responses in higher-order neurons, our results demonstrate that mouse auditory cortex can track fine frequency changes, which allows A2 Off responses in particular to better respond to pitch trajectories that distinguish behaviorally relevant, natural sound categories. SIGNIFICANCE STATEMENT A whistle's pitch conveys meaning to its listener, as when dogs learn that distinct pitch trajectories whistled by their owner differentiate specific commands. Many species use pitch trajectories in their own vocalizations to distinguish sound categories, such as in human languages, such as Mandarin. How and where auditory neural activity encodes these pitch trajectories as their meaning is learned but not well understood, especially for short-duration sounds. We studied this in mice, where infants use ultrasonic whistles to communicate to adults. We found that late neural firing after a sound ends can be tuned to how the pitch changes in time, and that this response in a secondary auditory cortical field changes with experience to acquire a pitch change's meaning.
Collapse
|
45
|
Metzen MG, Hofmann V, Chacron MJ. Neural Synchrony Gives Rise to Amplitude- and Duration-Invariant Encoding Consistent With Perception of Natural Communication Stimuli. Front Neurosci 2020; 14:79. [PMID: 32116522 PMCID: PMC7025533 DOI: 10.3389/fnins.2020.00079] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2019] [Accepted: 01/20/2020] [Indexed: 11/13/2022] Open
Abstract
When confronted with a highly variable environment, it remains poorly understood how neural populations encode and classify natural stimuli to give rise to appropriate and consistent behavioral responses. Here we investigated population coding of natural communication signals with different attributes (i.e., amplitude and duration) in the electrosensory system of the weakly electric fish Apteronotus leptorhynchus. Our results show that, while single peripheral neurons encode the detailed timecourse of different stimulus waveforms, measures of population synchrony are effectively unchanged because of coordinated increases and decreases in activity. A phenomenological mathematical model reproduced this invariance and shows that this can be explained by considering homogeneous populations whose responses are solely determined by single neuron firing properties. Moreover, recordings from downstream central neurons reveal that synchronous afferent activity is actually decoded and thus most likely transmitted to higher brain areas. Finally, we demonstrate that the associated behavioral responses at the organism level are invariant. Our results provide a mechanism by which amplitude- and duration-invariant coding of behaviorally relevant sensory input emerges across successive brain areas thereby presumably giving rise to invariant behavioral responses. Such mechanisms are likely to be found in other systems that share anatomical and functional features with the electrosensory system (e.g., auditory, visual, vestibular).
Collapse
Affiliation(s)
- Michael G Metzen
- Computational Systems Neuroscience Laboratory, Department of Physiology, McGill University, Montreal, QC, Canada
| | - Volker Hofmann
- Computational Systems Neuroscience Laboratory, Department of Physiology, McGill University, Montreal, QC, Canada
| | - Maurice J Chacron
- Computational Systems Neuroscience Laboratory, Department of Physiology, McGill University, Montreal, QC, Canada
| |
Collapse
|
46
|
Hosseini M, Rodriguez G, Guo H, Lim H, Plourde E. Novel metrics to measure the effect of additive inputs on the activity of sensory system neurons. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2019:5141-5145. [PMID: 31947016 DOI: 10.1109/embc.2019.8857622] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Sensory systems, such as the visual or auditory system, are highly non linear. It is therefore not easy to predict the effect of additive inputs on the spiking activity of related brain structures. Here, we propose two metrics to study the effect of additive covariates on the spiking activity of neurons. These metrics are directly obtained from a generalized linear model. We apply these metrics to the study of the effect of additive input audio noise on the spiking activity of neurons in the auditory system. To do so, we combine clean vocalisations with natural stationary or non-stationary noises and record activity in the auditory system while presenting the noisy vocalisations. We found that non-stationary noise has a greater effect on the neural activity than stationary noise. We observe that the results, obtained using the proposed metrics, is more consistent with current knowledge in auditory neuroscience than the results obtained when using a common metric from the literature, the extraction index.
Collapse
|
47
|
Experience- and Sex-Dependent Intrinsic Plasticity in the Zebra Finch Auditory Cortex during Song Memorization. J Neurosci 2020; 40:2047-2055. [PMID: 31937558 DOI: 10.1523/jneurosci.2137-19.2019] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2019] [Revised: 12/09/2019] [Accepted: 12/23/2019] [Indexed: 12/22/2022] Open
Abstract
For vocal communicators like humans and songbirds, survival and reproduction depend on highly developed auditory processing systems that can detect and differentiate nuanced differences in vocalizations, even amid noisy environments. Early auditory experience is critical to the development of these systems. In zebra finches and other songbirds, there is a sensitive period when young birds memorize a song that will serve as a model for their own vocal production. In addition to learning a specific tutor's song, the auditory system may also undergo critical developmental processes that support auditory perception of vocalizations more generally. Here, we investigate changes in intrinsic spiking dynamics among neurons in the caudal mesopallium, a cortical-level auditory area implicated in discriminating and learning species-specific vocalizations. A subset of neurons in this area only fire transiently at the onset of current injections (i.e., phasic firing), a dynamical property that can enhance the reliability and selectivity of neural responses to complex acoustic stimuli. At the beginning of the sensitive period, just after zebra finches have fledged from the nest, there is an increase in the proportion of caudal mesopallium neurons with phasic excitability, and in the proportion of neurons expressing Kv1.1, a low-threshold channel that facilitates phasic firing. This plasticity requires exposure to a complex, noisy environment and is greater in males, the only sex that sings in this species. This shift to more phasic dynamics is therefore an experience-dependent adaptation that could facilitate auditory processing in noisy, acoustically complex conditions during a key stage of vocal development.SIGNIFICANCE STATEMENT Auditory experience early in life shapes how humans and songbirds perceive the vocal communication sounds produced by their species. However, the changes that occur in the brain as this learning takes place are poorly understood. In this study, we show that in young zebra finches that are just beginning to learn the structure of their species' song, neurons in a key cortical area adapt their intrinsic firing patterns in response to the acoustic environment. In the complex, cocktail-party-like environment of a colony, more neurons adopt transient firing dynamics, which can facilitate neural coding of songs amid such challenging conditions.
Collapse
|
48
|
The Neuroethology of Vocal Communication in Songbirds: Production and Perception of a Call Repertoire. THE NEUROETHOLOGY OF BIRDSONG 2020. [DOI: 10.1007/978-3-030-34683-6_7] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
|
49
|
Woolley SC, Woolley SMN. Integrating Form and Function in the Songbird Auditory Forebrain. THE NEUROETHOLOGY OF BIRDSONG 2020. [DOI: 10.1007/978-3-030-34683-6_5] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
50
|
Ng CW, Recanzone GH. Age-Related Changes in Temporal Processing of Rapidly-Presented Sound Sequences in the Macaque Auditory Cortex. Cereb Cortex 2019; 28:3775-3796. [PMID: 29040403 DOI: 10.1093/cercor/bhx240] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2017] [Accepted: 08/31/2017] [Indexed: 11/13/2022] Open
Abstract
The mammalian auditory cortex is necessary to resolve temporal features in rapidly-changing sound streams. This capability is crucial for speech comprehension in humans and declines with normal aging. Nonhuman primate studies have revealed detrimental effects of normal aging on the auditory nervous system, and yet the underlying influence on temporal processing remains less well-defined. Therefore, we recorded from the core and lateral belt areas of auditory cortex when awake young and old monkeys listened to tone-pip and noise-burst sound sequences. Elevated spontaneous and stimulus-driven activity were the hallmark characteristics in old monkeys. These old neurons showed isomorphic-like discharge patterns to stimulus envelopes, though their phase-locking was less precise. Functional preference in temporal coding between the core and belt existed in the young monkeys but was mostly absent in the old monkeys, in which old belt neurons showed core-like response profiles. Finally, the analysis of population activity patterns indicated that the aged auditory cortex demonstrated a homogenous, distributed coding strategy, compared to the selective, sparse coding strategy observed in the young monkeys. Degraded temporal fidelity and highly-responsive, broadly-tuned cortical responses could underlie how aged humans have difficulties to resolve and track dynamic sounds leading to speech processing deficits.
Collapse
Affiliation(s)
- Chi-Wing Ng
- Center for the Neurobiology of Learning and Memory, University of California, Irvine, CA, USA
| | - Gregg H Recanzone
- Center for Neuroscience, University of California, Davis, CA, USA.,Department of Neurobiology, Physiology and Behavior, University of California, Davis, CA, USA
| |
Collapse
|