1
|
Fournier LA, Phadke RA, Salgado M, Brack A, Nocon JC, Bolshakova S, Grant JR, Padró Luna NM, Sen K, Cruz-Martín A. Overexpression of the schizophrenia risk gene C4 in PV cells drives sex-dependent behavioral deficits and circuit dysfunction. iScience 2024; 27:110800. [PMID: 39310747 PMCID: PMC11416532 DOI: 10.1016/j.isci.2024.110800] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 07/09/2024] [Accepted: 08/20/2024] [Indexed: 09/25/2024] Open
Abstract
Fast-spiking parvalbumin (PV)-positive cells are key players in orchestrating pyramidal neuron activity, and their dysfunction is consistently observed in myriad brain diseases. To understand how immune complement pathway dysregulation in PV cells drives disease pathogenesis, we have developed a transgenic line that permits cell-type specific overexpression of the schizophrenia-associated C4 gene. We found that overexpression of mouse C4 (mC4) in PV cells causes sex-specific alterations in anxiety-like behavior and deficits in synaptic connectivity and excitability of PFC PV cells. Using a computational model, we demonstrated that these microcircuit deficits led to hyperactivity and disrupted neural communication. Finally, pan-neuronal overexpression of mC4 failed to evoke the same deficits in behavior as PV-specific mC4 overexpression, suggesting that perturbations of this neuroimmune gene in fast-spiking neurons are especially detrimental to circuits associated with anxiety-like behavior. Together, these results provide a causative link between C4 and the vulnerability of PV cells in brain disease.
Collapse
Affiliation(s)
- Luke A. Fournier
- Neurobiology Section in the Department of Biology, Boston University, Boston, MA, USA
| | - Rhushikesh A. Phadke
- Molecular Biology, Cell Biology & Biochemistry Program, Boston University, Boston, MA, USA
| | - Maria Salgado
- Neurobiology Section in the Department of Biology, Boston University, Boston, MA, USA
| | - Alison Brack
- Molecular Biology, Cell Biology & Biochemistry Program, Boston University, Boston, MA, USA
| | - Jian Carlo Nocon
- Neurophotonics Center, Boston University, Boston, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, MA, USA
- Hearing Research Center, Boston University, Boston, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, MA, USA
| | - Sonia Bolshakova
- Neurobiology Section in the Department of Biology, Boston University, Boston, MA, USA
- Bioinformatics MS Program, Boston University, Boston, MA, USA
| | - Jaylyn R. Grant
- Biological Sciences, Eastern Illinois University, Charleston, IL, USA
- The Summer Undergraduate Research Fellowship (SURF) Program, Boston University, Boston, MA, USA
| | - Nicole M. Padró Luna
- The Summer Undergraduate Research Fellowship (SURF) Program, Boston University, Boston, MA, USA
- Biology Department, College of Natural Sciences, University of Puerto Rico, Rio Piedras Campus, San Juan, PR, USA
| | - Kamal Sen
- Neurophotonics Center, Boston University, Boston, MA, USA
- Center for Systems Neuroscience, Boston University, Boston, MA, USA
- Hearing Research Center, Boston University, Boston, MA, USA
- Department of Biomedical Engineering, Boston University, Boston, MA, USA
| | - Alberto Cruz-Martín
- Neurobiology Section in the Department of Biology, Boston University, Boston, MA, USA
- Molecular Biology, Cell Biology & Biochemistry Program, Boston University, Boston, MA, USA
- Department of Anesthesiology, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
- NeuroTechnology Center (NTC), University of Colorado Anschutz Medical Campus, Aurora, CO, USA
- Neuroscience Graduate Program, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
| |
Collapse
|
2
|
Gulati D, Ray S. Auditory and Visual Gratings Elicit Distinct Gamma Responses. eNeuro 2024; 11:ENEURO.0116-24.2024. [PMID: 38604776 PMCID: PMC11046261 DOI: 10.1523/eneuro.0116-24.2024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2024] [Revised: 04/01/2024] [Accepted: 04/03/2024] [Indexed: 04/13/2024] Open
Abstract
Sensory stimulation is often accompanied by fluctuations at high frequencies (>30 Hz) in brain signals. These could be "narrowband" oscillations in the gamma band (30-70 Hz) or nonoscillatory "broadband" high-gamma (70-150 Hz) activity. Narrowband gamma oscillations, which are induced by presenting some visual stimuli such as gratings and have been shown to weaken with healthy aging and the onset of Alzheimer's disease, hold promise as potential biomarkers. However, since delivering visual stimuli is cumbersome as it requires head stabilization for eye tracking, an equivalent auditory paradigm could be useful. Although simple auditory stimuli have been shown to produce high-gamma activity, whether specific auditory stimuli can also produce narrowband gamma oscillations is unknown. We tested whether auditory ripple stimuli, which are considered an analog to visual gratings, could elicit narrowband oscillations in auditory areas. We recorded 64-channel electroencephalogram from male and female (18 each) subjects while they either fixated on the monitor while passively viewing static visual gratings or listened to stationary and moving ripples, played using loudspeakers, with their eyes open or closed. We found that while visual gratings induced narrowband gamma oscillations with suppression in the alpha band (8-12 Hz), auditory ripples did not produce narrowband gamma but instead elicited very strong broadband high-gamma response and suppression in the beta band (14-26 Hz). Even though we used equivalent stimuli in both modalities, our findings indicate that the underlying neuronal circuitry may not share ubiquitous strategies for stimulus processing.
Collapse
Affiliation(s)
- Divya Gulati
- Centre for Neuroscience, Indian Institute of Science, Bengaluru 560012, India
| | - Supratim Ray
- Centre for Neuroscience, Indian Institute of Science, Bengaluru 560012, India
| |
Collapse
|
3
|
Khoury CF, Fala NG, Runyan CA. Arousal and Locomotion Differently Modulate Activity of Somatostatin Neurons across Cortex. eNeuro 2023; 10:ENEURO.0136-23.2023. [PMID: 37169583 PMCID: PMC10216262 DOI: 10.1523/eneuro.0136-23.2023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 04/28/2023] [Accepted: 05/03/2023] [Indexed: 05/13/2023] Open
Abstract
Arousal powerfully influences cortical activity, in part by modulating local inhibitory circuits. Somatostatin (SOM)-expressing inhibitory interneurons are particularly well situated to shape local population activity in response to shifts in arousal, yet the relationship between arousal state and SOM activity has not been characterized outside of sensory cortex. To determine whether SOM activity is similarly modulated by behavioral state across different levels of the cortical processing hierarchy, we compared the behavioral modulation of SOM-expressing neurons in auditory cortex (AC), a primary sensory region, and posterior parietal cortex (PPC), an association-level region of cortex, in mice. Behavioral state modulated activity differently in AC and PPC. In PPC, transitions to high arousal were accompanied by large increases in activity across the full PPC neural population, especially in SOM neurons. In AC, arousal transitions led to more subtle changes in overall activity, as individual SOM and Non-SOM neurons could be either positively or negatively modulated during transitions to high arousal states. The coding of sensory information in population activity was enhanced during periods of high arousal in AC, but not in PPC. Our findings suggest unique relationships between activity in local circuits and arousal across cortex, which may be tailored to the roles of specific cortical regions in sensory processing or the control of behavior.
Collapse
Affiliation(s)
- Christine F Khoury
- Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, Pennsylvania 15213
- Department of Neuroscience, University of Pittsburgh, Pittsburgh, Pennsylvania 15260
| | - Noelle G Fala
- Department of Neuroscience, University of Pittsburgh, Pittsburgh, Pennsylvania 15260
| | - Caroline A Runyan
- Center for the Neural Basis of Cognition, University of Pittsburgh, Pittsburgh, Pennsylvania 15213
- Department of Neuroscience, University of Pittsburgh, Pittsburgh, Pennsylvania 15260
| |
Collapse
|
4
|
Xu N, Zhao B, Luo L, Zhang K, Shao X, Luan G, Wang Q, Hu W, Wang Q. Two stages of speech envelope tracking in human auditory cortex modulated by speech intelligibility. Cereb Cortex 2023; 33:2215-2228. [PMID: 35695785 DOI: 10.1093/cercor/bhac203] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2022] [Revised: 05/01/2022] [Accepted: 05/02/2022] [Indexed: 11/13/2022] Open
Abstract
The envelope is essential for speech perception. Recent studies have shown that cortical activity can track the acoustic envelope. However, whether the tracking strength reflects the extent of speech intelligibility processing remains controversial. Here, using stereo-electroencephalogram technology, we directly recorded the activity in human auditory cortex while subjects listened to either natural or noise-vocoded speech. These 2 stimuli have approximately identical envelopes, but the noise-vocoded speech does not have speech intelligibility. According to the tracking lags, we revealed 2 stages of envelope tracking: an early high-γ (60-140 Hz) power stage that preferred the noise-vocoded speech and a late θ (4-8 Hz) phase stage that preferred the natural speech. Furthermore, the decoding performance of high-γ power was better in primary auditory cortex than in nonprimary auditory cortex, consistent with its short tracking delay, while θ phase showed better decoding performance in right auditory cortex. In addition, high-γ responses with sustained temporal profiles in nonprimary auditory cortex were dominant in both envelope tracking and decoding. In sum, we suggested a functional dissociation between high-γ power and θ phase: the former reflects fast and automatic processing of brief acoustic features, while the latter correlates to slow build-up processing facilitated by speech intelligibility.
Collapse
Affiliation(s)
- Na Xu
- Department of Neurology, Beijing Tiantan Hospital, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China.,National Clinical Research Center for Neurological Diseases, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China
| | - Baotian Zhao
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China
| | - Lu Luo
- School of Psychology, Beijing Sport University, No. 48 Xinxi Road, Haidian District, Beijing 100084, China
| | - Kai Zhang
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China
| | - Xiaoqiu Shao
- Department of Neurology, Beijing Tiantan Hospital, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China
| | - Guoming Luan
- Beijing Key Laboratory of Epilepsy, Epilepsy Center, Sanbo Brain Hospital, Capital Medical University, No. 50 Yikesong Xiangshan Road, Haidian District, Beijing 100093, China.,Beijing Institute of Brain Disorders, Collaborative Innovation Center for Brain Disorders, Capital Medical University, No.10 Xitoutiao, You An Men, Beijing 100069, China
| | - Qian Wang
- Beijing Key Laboratory of Epilepsy, Epilepsy Center, Sanbo Brain Hospital, Capital Medical University, No. 50 Yikesong Xiangshan Road, Haidian District, Beijing 100093, China.,School of Psychological and Cognitive Sciences, Beijing Key Laboratory of Behavior and Mental Health, Peking University, No.5 Yiheyuan Road, Haidian District, Beijing 100871, China.,IDG/McGovern Institute for Brain Research, Peking University, No.5 Yiheyuan Road, Haidian District, Beijing 100871, China
| | - Wenhan Hu
- Beijing Neurosurgical Institute, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China
| | - Qun Wang
- Department of Neurology, Beijing Tiantan Hospital, Capital Medical University, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China.,National Clinical Research Center for Neurological Diseases, No. 119 South Fourth Ring West Road, Fengtai District, Beijing 100070, China.,Beijing Institute of Brain Disorders, Collaborative Innovation Center for Brain Disorders, Capital Medical University, No.10 Xitoutiao, You An Men, Beijing 100069, China
| |
Collapse
|
5
|
Simon JZ, Commuri V, Kulasingham JP. Time-locked auditory cortical responses in the high-gamma band: A window into primary auditory cortex. Front Neurosci 2022; 16:1075369. [PMID: 36570848 PMCID: PMC9773383 DOI: 10.3389/fnins.2022.1075369] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 11/24/2022] [Indexed: 12/13/2022] Open
Abstract
Primary auditory cortex is a critical stage in the human auditory pathway, a gateway between subcortical and higher-level cortical areas. Receiving the output of all subcortical processing, it sends its output on to higher-level cortex. Non-invasive physiological recordings of primary auditory cortex using electroencephalography (EEG) and magnetoencephalography (MEG), however, may not have sufficient specificity to separate responses generated in primary auditory cortex from those generated in underlying subcortical areas or neighboring cortical areas. This limitation is important for investigations of effects of top-down processing (e.g., selective-attention-based) on primary auditory cortex: higher-level areas are known to be strongly influenced by top-down processes, but subcortical areas are often assumed to perform strictly bottom-up processing. Fortunately, recent advances have made it easier to isolate the neural activity of primary auditory cortex from other areas. In this perspective, we focus on time-locked responses to stimulus features in the high gamma band (70-150 Hz) and with early cortical latency (∼40 ms), intermediate between subcortical and higher-level areas. We review recent findings from physiological studies employing either repeated simple sounds or continuous speech, obtaining either a frequency following response (FFR) or temporal response function (TRF). The potential roles of top-down processing are underscored, and comparisons with invasive intracranial EEG (iEEG) and animal model recordings are made. We argue that MEG studies employing continuous speech stimuli may offer particular benefits, in that only a few minutes of speech generates robust high gamma responses from bilateral primary auditory cortex, and without measurable interference from subcortical or higher-level areas.
Collapse
Affiliation(s)
- Jonathan Z. Simon
- Department of Electrical and Computer Engineering, University of Maryland, College Park, College Park, MD, United States
- Department of Biology, University of Maryland, College Park, College Park, MD, United States
- Institute for Systems Research, University of Maryland, College Park, College Park, MD, United States
| | - Vrishab Commuri
- Department of Electrical and Computer Engineering, University of Maryland, College Park, College Park, MD, United States
| | | |
Collapse
|
6
|
Bonetti L, Brattico E, Bruzzone SEP, Donati G, Deco G, Pantazis D, Vuust P, Kringelbach ML. Brain recognition of previously learned versus novel temporal sequences: a differential simultaneous processing. Cereb Cortex 2022; 33:5524-5537. [PMID: 36346308 PMCID: PMC10152090 DOI: 10.1093/cercor/bhac439] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2022] [Revised: 10/12/2022] [Accepted: 12/13/2022] [Indexed: 11/09/2022] Open
Abstract
Abstract
Memory for sequences is a central topic in neuroscience, and decades of studies have investigated the neural mechanisms underlying the coding of a wide array of sequences extended over time. Yet, little is known on the brain mechanisms underlying the recognition of previously memorized versus novel temporal sequences. Moreover, the differential brain processing of single items in an auditory temporal sequence compared to the whole superordinate sequence is not fully understood. In this magnetoencephalography (MEG) study, the items of the temporal sequence were independently linked to local and rapid (2–8 Hz) brain processing, while the whole sequence was associated with concurrent global and slower (0.1–1 Hz) processing involving a widespread network of sequentially active brain regions. Notably, the recognition of previously memorized temporal sequences was associated to stronger activity in the slow brain processing, while the novel sequences required a greater involvement of the faster brain processing. Overall, the results expand on well-known information flow from lower- to higher order brain regions. In fact, they reveal the differential involvement of slow and faster whole brain processing to recognize previously learned versus novel temporal information.
Collapse
Affiliation(s)
- L Bonetti
- Center for Music in the Brain (MIB), Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg , Universitetsbyen 3, 8000, Aarhus C , Denmark
- Centre for Eudaimonia and Human Flourishing, Linacre College, University of Oxford , Stoke place 7, OX39BX, Oxford , UK
- University of Oxford Department of Psychiatry, , Oxford, UK
- University of Bologna Department of Psychology, , Italy
| | - E Brattico
- Center for Music in the Brain (MIB), Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg , Universitetsbyen 3, 8000, Aarhus C , Denmark
- University of Bari Aldo Moro Department of Education, Psychology, Communication, , Italy
| | - S E P Bruzzone
- Center for Music in the Brain (MIB) , Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Universitetsbyen 3, 8000, Aarhus C , Denmark
- Copenhagen University Hospital Rigshospitalet Neurobiology Research Unit (NRU), , Inge Lehmanns Vej 6, 2100, Copenhagen , Denmark
- Faculty of Health and Medical Sciences, University of Copenhagen , Blegdamsvej 3B, 2200, Copenhagen , Denmark
| | - G Donati
- University of Bologna Department of Psychology, , Italy
| | - G Deco
- Center for Brain and Cognition, Universitat Pompeu Fabra Computational and Theoretical Neuroscience Group, , Edifici Merce Rodereda, C/ de Ramon Trias Fargas, 25, 08018 Barcelona , Spain
| | - D Pantazis
- McGovern Institute for Brain Research, Massachusetts Institute of Technology (MIT) , 77 Massachusetts Ave, Cambridge, MA 02139 , USA
| | - P Vuust
- Center for Music in the Brain (MIB), Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg , Universitetsbyen 3, 8000, Aarhus C , Denmark
| | - M L Kringelbach
- Center for Music in the Brain (MIB), Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg , Universitetsbyen 3, 8000, Aarhus C , Denmark
- Centre for Eudaimonia and Human Flourishing, Linacre College, University of Oxford , Stoke place 7, OX39BX, Oxford , UK
- University of Oxford Department of Psychiatry, , Oxford, UK
| |
Collapse
|
7
|
Khoury CF, Fala NG, Runyan CA. The spatial scale of somatostatin subnetworks increases from sensory to association cortex. Cell Rep 2022; 40:111319. [PMID: 36070697 DOI: 10.1016/j.celrep.2022.111319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2022] [Revised: 07/01/2022] [Accepted: 08/15/2022] [Indexed: 11/24/2022] Open
Abstract
Incoming signals interact with rich, ongoing population activity dynamics in cortical circuits. These intrinsic dynamics are the consequence of interactions among local excitatory and inhibitory neurons and affect inter-region communication and information coding. It is unclear whether specializations in the patterns of interactions among excitatory and inhibitory neurons underlie systematic differences in activity dynamics across the cortex. Here, in mice, we compare the functional interactions among somatostatin (SOM)-expressing inhibitory interneurons and the rest of the neural population in auditory cortex (AC), a sensory region of the cortex, and posterior parietal cortex (PPC), an association region. The spatial structure of shared variability among SOM and non-SOM neurons differs across regions: correlations decay rapidly with distance in AC but not in PPC. However, in both regions, activity of SOM neurons is more highly correlated than non-SOM neurons' activity. Our results imply both generalization and specialization in the functional structure of inhibitory subnetworks across the cortex.
Collapse
Affiliation(s)
- Christine F Khoury
- Center for the Neural Basis of Cognition, Pittsburgh, PA 15213, USA; Department of Neuroscience, University of Pittsburgh, Pittsburgh, PA 15260, USA
| | - Noelle G Fala
- Department of Neuroscience, University of Pittsburgh, Pittsburgh, PA 15260, USA
| | - Caroline A Runyan
- Center for the Neural Basis of Cognition, Pittsburgh, PA 15213, USA; Department of Neuroscience, University of Pittsburgh, Pittsburgh, PA 15260, USA.
| |
Collapse
|
8
|
Echolocation-related reversal of information flow in a cortical vocalization network. Nat Commun 2022; 13:3642. [PMID: 35752629 PMCID: PMC9233670 DOI: 10.1038/s41467-022-31230-6] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 05/30/2022] [Indexed: 11/09/2022] Open
Abstract
The mammalian frontal and auditory cortices are important for vocal behavior. Here, using local-field potential recordings, we demonstrate that the timing and spatial patterns of oscillations in the fronto-auditory network of vocalizing bats (Carollia perspicillata) predict the purpose of vocalization: echolocation or communication. Transfer entropy analyses revealed predominant top-down (frontal-to-auditory cortex) information flow during spontaneous activity and pre-vocal periods. The dynamics of information flow depend on the behavioral role of the vocalization and on the timing relative to vocal onset. We observed the emergence of predominant bottom-up (auditory-to-frontal) information transfer during the post-vocal period specific to echolocation pulse emission, leading to self-directed acoustic feedback. Electrical stimulation of frontal areas selectively enhanced responses to sounds in auditory cortex. These results reveal unique changes in information flow across sensory and frontal cortices, potentially driven by the purpose of the vocalization in a highly vocal mammalian model.
Collapse
|
9
|
Brinkman BAW, Yan H, Maffei A, Park IM, Fontanini A, Wang J, La Camera G. Metastable dynamics of neural circuits and networks. APPLIED PHYSICS REVIEWS 2022; 9:011313. [PMID: 35284030 PMCID: PMC8900181 DOI: 10.1063/5.0062603] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/06/2021] [Accepted: 01/31/2022] [Indexed: 05/14/2023]
Abstract
Cortical neurons emit seemingly erratic trains of action potentials or "spikes," and neural network dynamics emerge from the coordinated spiking activity within neural circuits. These rich dynamics manifest themselves in a variety of patterns, which emerge spontaneously or in response to incoming activity produced by sensory inputs. In this Review, we focus on neural dynamics that is best understood as a sequence of repeated activations of a number of discrete hidden states. These transiently occupied states are termed "metastable" and have been linked to important sensory and cognitive functions. In the rodent gustatory cortex, for instance, metastable dynamics have been associated with stimulus coding, with states of expectation, and with decision making. In frontal, parietal, and motor areas of macaques, metastable activity has been related to behavioral performance, choice behavior, task difficulty, and attention. In this article, we review the experimental evidence for neural metastable dynamics together with theoretical approaches to the study of metastable activity in neural circuits. These approaches include (i) a theoretical framework based on non-equilibrium statistical physics for network dynamics; (ii) statistical approaches to extract information about metastable states from a variety of neural signals; and (iii) recent neural network approaches, informed by experimental results, to model the emergence of metastable dynamics. By discussing these topics, we aim to provide a cohesive view of how transitions between different states of activity may provide the neural underpinnings for essential functions such as perception, memory, expectation, or decision making, and more generally, how the study of metastable neural activity may advance our understanding of neural circuit function in health and disease.
Collapse
Affiliation(s)
| | - H. Yan
- State Key Laboratory of Electroanalytical Chemistry, Changchun Institute of Applied Chemistry, Chinese Academy of Sciences, Changchun, Jilin 130022, People's Republic of China
| | | | | | | | - J. Wang
- Authors to whom correspondence should be addressed: and
| | - G. La Camera
- Authors to whom correspondence should be addressed: and
| |
Collapse
|
10
|
Nishimura M, Song WJ. Region-dependent Millisecond Time-scale Sensitivity in Spectrotemporal Integrations in Guinea Pig Primary Auditory Cortex. Neuroscience 2022; 480:229-245. [PMID: 34762984 DOI: 10.1016/j.neuroscience.2021.10.030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2021] [Revised: 10/28/2021] [Accepted: 10/29/2021] [Indexed: 11/18/2022]
Abstract
Spectrotemporal integration is a key function of our auditory system for discriminating spectrotemporally complex sounds, such as words. Response latency in the auditory cortex is known to change with the millisecond time-scale depending on acoustic parameters, such as sound frequency and intensity. The functional significance of the millisecond-range latency difference in the integration remains unclear. Actually, whether the auditory cortex has a sensitivity to the millisecond-range difference has not been systematically examined. Herein, we examined the sensitivity in the primary auditory cortex (A1) using voltage-sensitive dye imaging techniques in guinea pigs. Bandpass noise bursts in two different bands (band-noises), centered at 1 and 16 kHz, respectively, were used for the examination. Onset times of individual band-noises (spectral onset-times) were varied to virtually cancel or magnify the latency difference observed with the band-noises. Conventionally defined nonlinear effects in integration were analyzed at A1 with varying sound intensities (or response latencies) and/or spectral onset-times of the two band-noises. The nonlinear effect measured in the high-frequency region of the A1 linearly changed depending on the millisecond difference of the response onset-times, which were estimated from the spatially-local response latencies and spectral onset-times. In contrast, the low-frequency region of the A1 had no significant sensitivity to the millisecond difference. The millisecond-range latency difference may have functional significance in the spectrotemporal integration with the millisecond time-scale sensitivity at the high-frequency region of A1 but not at the low-frequency region.
Collapse
Affiliation(s)
- Masataka Nishimura
- Department of Sensory and Cognitive Physiology, Faculty of Life Sciences, Kumamoto University, 1-1-1 Honjo, Kumamoto 8608556, Japan.
| | - Wen-Jie Song
- Department of Sensory and Cognitive Physiology, Faculty of Life Sciences, Kumamoto University, 1-1-1 Honjo, Kumamoto 8608556, Japan; Program for Leading Graduate Schools HIGO Program, Kumamoto University, Kumamoto, Japan
| |
Collapse
|
11
|
Bonetti L, Brattico E, Carlomagno F, Donati G, Cabral J, Haumann NT, Deco G, Vuust P, Kringelbach ML. Rapid encoding of musical tones discovered in whole-brain connectivity. Neuroimage 2021; 245:118735. [PMID: 34813972 DOI: 10.1016/j.neuroimage.2021.118735] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2021] [Revised: 09/30/2021] [Accepted: 11/14/2021] [Indexed: 11/26/2022] Open
Abstract
Information encoding has received a wide neuroscientific attention, but the underlying rapid spatiotemporal brain dynamics remain largely unknown. Here, we investigated the rapid brain mechanisms for encoding of sounds forming a complex temporal sequence. Specifically, we used magnetoencephalography (MEG) to record the brain activity of 68 participants while they listened to a highly structured musical prelude. Functional connectivity analyses performed using phase synchronisation and graph theoretical measures showed a large network of brain areas recruited during encoding of sounds, comprising primary and secondary auditory cortices, frontal operculum, insula, hippocampus and basal ganglia. Moreover, our results highlighted the rapid transition of brain activity from primary auditory cortex to higher order association areas including insula and superior temporal pole within a whole-brain network, occurring during the first 220 ms of the encoding process. Further, we discovered that individual differences along cognitive abilities and musicianship modulated the degree centrality of the brain areas implicated in the encoding process. Indeed, participants with higher musical expertise presented a stronger centrality of superior temporal gyrus and insula, while individuals with high working memory abilities showed a stronger centrality of frontal operculum. In conclusion, our study revealed the rapid unfolding of brain network dynamics responsible for the encoding of sounds and their relationship with individual differences, showing a complex picture which extends beyond the well-known involvement of auditory areas. Indeed, our results expanded our understanding of the general mechanisms underlying auditory pattern encoding in the human brain.
Collapse
Affiliation(s)
- L Bonetti
- Centre for Eudaimonia and Human Flourishing, University of Oxford, United Kingdom; Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark; Department of Psychiatry, University of Oxford, Oxford, United Kingdom; Department of Psychology, University of Bologna, Italy.
| | - E Brattico
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark; Department of Education, Psychology, Communication, University of Bari Aldo Moro, Italy
| | - F Carlomagno
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark
| | - G Donati
- Department of Psychology, University of Bologna, Italy; Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark
| | - J Cabral
- Centre for Eudaimonia and Human Flourishing, University of Oxford, United Kingdom; Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark; Life and Health Sciences Research Institute (ICVS), School of Medicine, University of Minho, 4710-057 Braga, Portugal
| | - N T Haumann
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark
| | - G Deco
- Institució Catalana de la Recerca i Estudis Avançats (ICREA), Passeig Lluís Companys 23, Barcelona, 08010, Spain; Computational and Theoretical Neuroscience Group, Center for Brain and Cognition, Universitat Pompeu Fabra, Barcelona, Spain
| | - P Vuust
- Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark
| | - M L Kringelbach
- Centre for Eudaimonia and Human Flourishing, University of Oxford, United Kingdom; Center for Music in the Brain, Department of Clinical Medicine, Aarhus University & The Royal Academy of Music Aarhus/Aalborg, Denmark; Department of Psychiatry, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
12
|
Mathematical framework for place coding in the auditory system. PLoS Comput Biol 2021; 17:e1009251. [PMID: 34339409 PMCID: PMC8360601 DOI: 10.1371/journal.pcbi.1009251] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2020] [Revised: 08/12/2021] [Accepted: 07/06/2021] [Indexed: 11/18/2022] Open
Abstract
In the auditory system, tonotopy is postulated to be the substrate for a place code, where sound frequency is encoded by the location of the neurons that fire during the stimulus. Though conceptually simple, the computations that allow for the representation of intensity and complex sounds are poorly understood. Here, a mathematical framework is developed in order to define clearly the conditions that support a place code. To accommodate both frequency and intensity information, the neural network is described as a space with elements that represent individual neurons and clusters of neurons. A mapping is then constructed from acoustic space to neural space so that frequency and intensity are encoded, respectively, by the location and size of the clusters. Algebraic operations -addition and multiplication- are derived to elucidate the rules for representing, assembling, and modulating multi-frequency sound in networks. The resulting outcomes of these operations are consistent with network simulations as well as with electrophysiological and psychophysical data. The analyses show how both frequency and intensity can be encoded with a purely place code, without the need for rate or temporal coding schemes. The algebraic operations are used to describe loudness summation and suggest a mechanism for the critical band. The mathematical approach complements experimental and computational approaches and provides a foundation for interpreting data and constructing models.
Collapse
|
13
|
Homma NY, Atencio CA, Schreiner CE. Plasticity of Multidimensional Receptive Fields in Core Rat Auditory Cortex Directed by Sound Statistics. Neuroscience 2021; 467:150-170. [PMID: 33951506 DOI: 10.1016/j.neuroscience.2021.04.028] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2020] [Revised: 04/09/2021] [Accepted: 04/24/2021] [Indexed: 11/17/2022]
Abstract
Sensory cortical neurons can nonlinearly integrate a wide range of inputs. The outcome of this nonlinear process can be approximated by more than one receptive field component or filter to characterize the ensuing stimulus preference. The functional properties of multidimensional filters are, however, not well understood. Here we estimated two spectrotemporal receptive fields (STRFs) per neuron using maximally informative dimension analysis. We compared their temporal and spectral modulation properties and determined the stimulus information captured by the two STRFs in core rat auditory cortical fields, primary auditory cortex (A1) and ventral auditory field (VAF). The first STRF is the dominant filter and acts as a sound feature detector in both fields. The second STRF is less feature specific, preferred lower modulations, and had less spike information compared to the first STRF. The information jointly captured by the two STRFs was larger than that captured by the sum of the individual STRFs, reflecting nonlinear interactions of two filters. This information gain was larger in A1. We next determined how the acoustic environment affects the structure and relationship of these two STRFs. Rats were exposed to moderate levels of spectrotemporally modulated noise during development. Noise exposure strongly altered the spectrotemporal preference of the first STRF in both cortical fields. The interaction between the two STRFs was reduced by noise exposure in A1 but not in VAF. The results reveal new functional distinctions between A1 and VAF indicating that (i) A1 has stronger interactions of the two STRFs than VAF, (ii) noise exposure diminishes modulation parameter representation contained in the noise more strongly for the first STRF in both fields, and (iii) plasticity induced by noise exposure can affect the strength of filter interactions in A1. Taken together, ascertaining two STRFs per neuron enhances the understanding of cortical information processing and plasticity effects in core auditory cortex.
Collapse
Affiliation(s)
- Natsumi Y Homma
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California San Francisco, San Francisco, USA; Center for Integrative Neuroscience, University of California San Francisco, San Francisco, USA.
| | - Craig A Atencio
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California San Francisco, San Francisco, USA
| | - Christoph E Schreiner
- Coleman Memorial Laboratory, Department of Otolaryngology - Head and Neck Surgery, University of California San Francisco, San Francisco, USA; Center for Integrative Neuroscience, University of California San Francisco, San Francisco, USA
| |
Collapse
|
14
|
Kulasingham JP, Brodbeck C, Presacco A, Kuchinsky SE, Anderson S, Simon JZ. High gamma cortical processing of continuous speech in younger and older listeners. Neuroimage 2020; 222:117291. [PMID: 32835821 PMCID: PMC7736126 DOI: 10.1016/j.neuroimage.2020.117291] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2020] [Revised: 08/12/2020] [Accepted: 08/16/2020] [Indexed: 12/11/2022] Open
Abstract
Neural processing along the ascending auditory pathway is often associated with a progressive reduction in characteristic processing rates. For instance, the well-known frequency-following response (FFR) of the auditory midbrain, as measured with electroencephalography (EEG), is dominated by frequencies from ∼100 Hz to several hundred Hz, phase-locking to the acoustic stimulus at those frequencies. In contrast, cortical responses, whether measured by EEG or magnetoencephalography (MEG), are typically characterized by frequencies of a few Hz to a few tens of Hz, time-locking to acoustic envelope features. In this study we investigated a crossover case, cortically generated responses time-locked to continuous speech features at FFR-like rates. Using MEG, we analyzed responses in the high gamma range of 70-200 Hz to continuous speech using neural source-localized reverse correlation and the corresponding temporal response functions (TRFs). Continuous speech stimuli were presented to 40 subjects (17 younger, 23 older adults) with clinically normal hearing and their MEG responses were analyzed in the 70-200 Hz band. Consistent with the relative insensitivity of MEG to many subcortical structures, the spatiotemporal profile of these response components indicated a cortical origin with ∼40 ms peak latency and a right hemisphere bias. TRF analysis was performed using two separate aspects of the speech stimuli: a) the 70-200 Hz carrier of the speech, and b) the 70-200 Hz temporal modulations in the spectral envelope of the speech stimulus. The response was dominantly driven by the envelope modulation, with a much weaker contribution from the carrier. Age-related differences were also analyzed to investigate a reversal previously seen along the ascending auditory pathway, whereby older listeners show weaker midbrain FFR responses than younger listeners, but, paradoxically, have stronger cortical low frequency responses. In contrast to both these earlier results, this study did not find clear age-related differences in high gamma cortical responses to continuous speech. Cortical responses at FFR-like frequencies shared some properties with midbrain responses at the same frequencies and with cortical responses at much lower frequencies.
Collapse
Affiliation(s)
- Joshua P Kulasingham
- (a)Department of Electrical and Computer Engineering, University of Maryland, College Park, MD, United States.
| | - Christian Brodbeck
- (b)Institute for Systems Research, University of Maryland, College Park, Maryland, United States.
| | - Alessandro Presacco
- (b)Institute for Systems Research, University of Maryland, College Park, Maryland, United States.
| | - Stefanie E Kuchinsky
- (c)Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, Maryland, United States.
| | - Samira Anderson
- (d)Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland, United States.
| | - Jonathan Z Simon
- (a)Department of Electrical and Computer Engineering, University of Maryland, College Park, MD, United States; (b)Institute for Systems Research, University of Maryland, College Park, Maryland, United States; (e)Department of Biology, University of Maryland, College Park, Maryland, United States.
| |
Collapse
|
15
|
Baroni F, Morillon B, Trébuchon A, Liégeois-Chauvel C, Olasagasti I, Giraud AL. Converging intracortical signatures of two separated processing timescales in human early auditory cortex. Neuroimage 2020; 218:116882. [PMID: 32439539 DOI: 10.1016/j.neuroimage.2020.116882] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2019] [Revised: 03/30/2020] [Accepted: 04/23/2020] [Indexed: 11/15/2022] Open
Abstract
Neural oscillations in auditory cortex are argued to support parsing and representing speech constituents at their corresponding temporal scales. Yet, how incoming sensory information interacts with ongoing spontaneous brain activity, what features of the neuronal microcircuitry underlie spontaneous and stimulus-evoked spectral fingerprints, and what these fingerprints entail for stimulus encoding, remain largely open questions. We used a combination of human invasive electrophysiology, computational modeling and decoding techniques to assess the information encoding properties of brain activity and to relate them to a plausible underlying neuronal microarchitecture. We analyzed intracortical auditory EEG activity from 10 patients while they were listening to short sentences. Pre-stimulus neural activity in early auditory cortical regions often exhibited power spectra with a shoulder in the delta range and a small bump in the beta range. Speech decreased power in the beta range, and increased power in the delta-theta and gamma ranges. Using multivariate machine learning techniques, we assessed the spectral profile of information content for two aspects of speech processing: detection and discrimination. We obtained better phase than power information decoding, and a bimodal spectral profile of information content with better decoding at low (delta-theta) and high (gamma) frequencies than at intermediate (beta) frequencies. These experimental data were reproduced by a simple rate model made of two subnetworks with different timescales, each composed of coupled excitatory and inhibitory units, and connected via a negative feedback loop. Modeling and experimental results were similar in terms of pre-stimulus spectral profile (except for the iEEG beta bump), spectral modulations with speech, and spectral profile of information content. Altogether, we provide converging evidence from both univariate spectral analysis and decoding approaches for a dual timescale processing infrastructure in human auditory cortex, and show that it is consistent with the dynamics of a simple rate model.
Collapse
Affiliation(s)
- Fabiano Baroni
- Department of Fundamental Neuroscience, University of Geneva, Geneva, Switzerland; School of Engineering, École Polytechnique Fédérale de Lausanne, Lausanne, Switzerland.
| | - Benjamin Morillon
- Aix Marseille Université, Institut National de la Santé et de la Recherche Médicale (INSERM), Institut de Neurosciences des Systémes (INS), Marseille, France
| | - Agnès Trébuchon
- Aix Marseille Université, Institut National de la Santé et de la Recherche Médicale (INSERM), Institut de Neurosciences des Systémes (INS), Marseille, France; Clinical Neurophysiology and Epileptology Department, Timone Hospital, Assistance Publique Hôpitaux de Marseille, Marseille, France
| | - Catherine Liégeois-Chauvel
- Aix Marseille Université, Institut National de la Santé et de la Recherche Médicale (INSERM), Institut de Neurosciences des Systémes (INS), Marseille, France; Department of Neurological Surgery, University of Pittsburgh, PA, 15213, USA
| | - Itsaso Olasagasti
- Department of Fundamental Neuroscience, University of Geneva, Geneva, Switzerland
| | - Anne-Lise Giraud
- Department of Fundamental Neuroscience, University of Geneva, Geneva, Switzerland
| |
Collapse
|
16
|
Ross B, Tremblay KL, Alain C. Simultaneous EEG and MEG recordings reveal vocal pitch elicited cortical gamma oscillations in young and older adults. Neuroimage 2019; 204:116253. [PMID: 31600592 DOI: 10.1016/j.neuroimage.2019.116253] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2019] [Revised: 09/13/2019] [Accepted: 10/06/2019] [Indexed: 10/25/2022] Open
Abstract
The frequency-following response with origin in the auditory brainstem represents the pitch contour of voice and can be recorded with electrodes from the scalp. MEG studies also revealed a cortical contribution to the high gamma oscillations at the fundamental frequency (f0) of a vowel stimulus. Therefore, studying the cortical component of the frequency-following response could provide insights into how pitch information is encoded at the cortical level. Comparing how aging affects the different responses may help to uncover the neural mechanisms underlying speech understanding deficits in older age. We simultaneously recorded EEG and MEG responses to the syllable /ba/. MEG beamformer analysis localized sources in bilateral auditory cortices and the midbrain. Time-frequency analysis showed a faithful representation of the pitch contour between 106 Hz and 138 Hz in the cortical activity. A cross-correlation revealed a latency of 20 ms. Furthermore, stimulus onsets elicited cortical 40-Hz responses. Both the 40-Hz and the f0 response amplitudes increased in older age and were larger in the right hemisphere. The effects of aging and laterality of the f0 response were evident in the MEG only, suggesting that both effects were characteristics of the cortical response. After comparing f0 and N1 responses in EEG and MEG, we estimated that approximately one-third of the scalp-recorded f0 response could be cortical in origin. We attributed the significance of the cortical f0 response to the precise timing of cortical neurons that serve as a time-sensitive code for pitch.
Collapse
Affiliation(s)
- Bernhard Ross
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department for Medical Biophysics, University of Toronto, Ontario, Canada.
| | - Kelly L Tremblay
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada
| |
Collapse
|
17
|
Ismaail NM, Shalaby AA, Ibraheem OA. Effect of age on Gaps-In-Noise test in pediatric population. Int J Pediatr Otorhinolaryngol 2019; 122:155-160. [PMID: 31029950 DOI: 10.1016/j.ijporl.2019.04.010] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/28/2019] [Revised: 04/10/2019] [Accepted: 04/10/2019] [Indexed: 10/27/2022]
Abstract
OBJECTIVES The main objective was to examine the effect of central maturation on the auditory temporal resolution in a group of school-age children using Gaps-In-Noise test. METHODS The study involved 180 children (6-16 years) with normal hearing, average intelligence and language skills, and adequate scholastic achievement. Subjects were divided into four age subgroups. Investigations involved basic audiological evaluation, screening test battery for central auditory processing, and finally Gaps-In-Noise test. RESULTS Comparison of the four age subgroups revealed non-significant age effect on the Gaps-In-Noise test. The approximate gap detection threshold of children was comparable to that of adults. Equivalent data were obtained as a function of the ear, gender, list, and retest. CONCLUSION Central auditory maturation of the temporal resolution and hence the Gaps-In-Noise test has been established by age 5 years. Consequently, assessment of Gaps-In-Noise test in school-age children provided adult-like normative data. The stability of outcomes across different factors highlights the clinical validity of Gaps-In-Noise test in the assessment of temporal resolution deficit and follow-up after remediation.
Collapse
Affiliation(s)
- Naema M Ismaail
- Audio-vestibular Medicine Unit, Department of Otolaryngology-Head and Neck Surgery, Faculty of Medicine (Girls), University of AL-Azhar, Cairo, Egypt
| | - Amany A Shalaby
- Audio-vestibular Medicine Unit, Department of Otolaryngology-Head and Neck Surgery, Faculty of Medicine, University of Ain Shams, Cairo, Egypt
| | - Ola A Ibraheem
- Audio-vestibular Medicine Unit, Department of Otolaryngology-Head and Neck Surgery, Faculty of Medicine, University of Zagazig, Zagazig, Egypt.
| |
Collapse
|
18
|
Chen C, Read HL, Escabí MA. A temporal integration mechanism enhances frequency selectivity of broadband inputs to inferior colliculus. PLoS Biol 2019; 17:e2005861. [PMID: 31233489 PMCID: PMC6611646 DOI: 10.1371/journal.pbio.2005861] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2018] [Revised: 07/05/2019] [Accepted: 05/22/2019] [Indexed: 11/18/2022] Open
Abstract
Accurately resolving frequency components in sounds is essential for sound recognition, yet there is little direct evidence for how frequency selectivity is preserved or newly created across auditory structures. We demonstrate that prepotentials (PPs) with physiological properties resembling presynaptic potentials from broadly tuned brainstem inputs can be recorded concurrently with postsynaptic action potentials in inferior colliculus (IC). These putative brainstem inputs (PBIs) are broadly tuned and exhibit delayed and spectrally interleaved excitation and inhibition not present in the simultaneously recorded IC neurons (ICNs). A sharpening of tuning is accomplished locally at the expense of spike-timing precision through nonlinear temporal integration of broadband inputs. A neuron model replicates the finding and demonstrates that temporal integration alone can degrade timing precision while enhancing frequency tuning through interference of spectrally in- and out-of-phase inputs. These findings suggest that, in contrast to current models that require local inhibition, frequency selectivity can be sharpened through temporal integration, thus supporting an alternative computational strategy to quickly refine frequency selectivity.
Collapse
Affiliation(s)
- Chen Chen
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Heather L. Read
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| |
Collapse
|
19
|
Chambers JD, Elgueda D, Fritz JB, Shamma SA, Burkitt AN, Grayden DB. Computational Neural Modeling of Auditory Cortical Receptive Fields. Front Comput Neurosci 2019; 13:28. [PMID: 31178710 PMCID: PMC6543553 DOI: 10.3389/fncom.2019.00028] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2018] [Accepted: 04/23/2019] [Indexed: 11/13/2022] Open
Abstract
Previous studies have shown that the auditory cortex can enhance the perception of behaviorally important sounds in the presence of background noise, but the mechanisms by which it does this are not yet elucidated. Rapid plasticity of spectrotemporal receptive fields (STRFs) in the primary (A1) cortical neurons is observed during behavioral tasks that require discrimination of particular sounds. This rapid task-related change is believed to be one of the processing strategies utilized by the auditory cortex to selectively attend to one stream of sound in the presence of mixed sounds. However, the mechanism by which the brain evokes this rapid plasticity in the auditory cortex remains unclear. This paper uses a neural network model to investigate how synaptic transmission within the cortical neuron network can change the receptive fields of individual neurons. A sound signal was used as input to a model of the cochlea and auditory periphery, which activated or inhibited integrate-and-fire neuron models to represent networks in the primary auditory cortex. Each neuron in the network was tuned to a different frequency. All neurons were interconnected with excitatory or inhibitory synapses of varying strengths. Action potentials in one of the model neurons were used to calculate the receptive field using reverse correlation. The results were directly compared to previously recorded electrophysiological data from ferrets performing behavioral tasks that require discrimination of particular sounds. The neural network model could reproduce complex STRFs observed experimentally through optimizing the synaptic weights in the model. The model predicts that altering synaptic drive between cortical neurons and/or bottom-up synaptic drive from the cochlear model to the cortical neurons can account for rapid task-related changes observed experimentally in A1 neurons. By identifying changes in the synaptic drive during behavioral tasks, the model provides insights into the neural mechanisms utilized by the auditory cortex to enhance the perception of behaviorally salient sounds.
Collapse
Affiliation(s)
- Jordan D Chambers
- NeuroEngineering Laboratory, Department of Biomedical Engineering, University of Melbourne, Parkville, VIC, Australia
| | - Diego Elgueda
- Departamento de Patología Animal, Facultad de Ciencias Veterinarias y Pecuarias, Universidad de Chile, Santiago, Chile.,Institute for Systems Research, University of Maryland, College Park, MD, United States
| | - Jonathan B Fritz
- Institute for Systems Research, University of Maryland, College Park, MD, United States
| | - Shihab A Shamma
- Institute for Systems Research, University of Maryland, College Park, MD, United States.,Laboratoire des Systèmes Perceptifs, École Normale Supérieure, Paris, France
| | - Anthony N Burkitt
- NeuroEngineering Laboratory, Department of Biomedical Engineering, University of Melbourne, Parkville, VIC, Australia
| | - David B Grayden
- NeuroEngineering Laboratory, Department of Biomedical Engineering, University of Melbourne, Parkville, VIC, Australia
| |
Collapse
|
20
|
Krause BM, Murphy CA, Uhlrich DJ, Banks MI. PV+ Cells Enhance Temporal Population Codes but not Stimulus-Related Timing in Auditory Cortex. Cereb Cortex 2019; 29:627-647. [PMID: 29300837 PMCID: PMC6319178 DOI: 10.1093/cercor/bhx345] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2017] [Revised: 11/30/2017] [Accepted: 12/05/2017] [Indexed: 01/05/2023] Open
Abstract
Spatio-temporal cortical activity patterns relative to both peripheral input and local network activity carry information about stimulus identity and context. GABAergic interneurons are reported to regulate spiking at millisecond precision in response to sensory stimulation and during gamma oscillations; their role in regulating spike timing during induced network bursts is unclear. We investigated this issue in murine auditory thalamo-cortical (TC) brain slices, in which TC afferents induced network bursts similar to previous reports in vivo. Spike timing relative to TC afferent stimulation during bursts was poor in pyramidal cells and SOM+ interneurons. It was more precise in PV+ interneurons, consistent with their reported contribution to spiking precision in pyramidal cells. Optogenetic suppression of PV+ cells unexpectedly improved afferent-locked spike timing in pyramidal cells. In contrast, our evidence suggests that PV+ cells do regulate the spatio-temporal spike pattern of pyramidal cells during network bursts, whose organization is suited to ensemble coding of stimulus information. Simulations showed that suppressing PV+ cells reduces the capacity of pyramidal cell networks to produce discriminable spike patterns. By dissociating temporal precision with respect to a stimulus versus internal cortical activity, we identified a novel role for GABAergic cells in regulating information processing in cortical networks.
Collapse
Affiliation(s)
- Bryan M Krause
- Department of Anesthesiology, University of Wisconsin, Madison, WI, USA
- Neuroscience Training Program, University of Wisconsin, Madison, WI, USA
| | - Caitlin A Murphy
- Physiology Graduate Training Program, University of Wisconsin, Madison, WI, USA
| | - Daniel J Uhlrich
- Department of Neuroscience, University of Wisconsin, Madison, WI, USA
| | - Matthew I Banks
- Department of Anesthesiology, University of Wisconsin, Madison, WI, USA
| |
Collapse
|
21
|
García-Rosales F, Beetz MJ, Cabral-Calderin Y, Kössl M, Hechavarria JC. Neuronal coding of multiscale temporal features in communication sequences within the bat auditory cortex. Commun Biol 2018; 1:200. [PMID: 30480101 PMCID: PMC6244232 DOI: 10.1038/s42003-018-0205-5] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2018] [Accepted: 10/30/2018] [Indexed: 11/18/2022] Open
Abstract
Experimental evidence supports that cortical oscillations represent multiscale temporal modulations existent in natural stimuli, yet little is known about the processing of these multiple timescales at a neuronal level. Here, using extracellular recordings from the auditory cortex (AC) of awake bats (Carollia perspicillata), we show the existence of three neuronal types which represent different levels of the temporal structure of conspecific vocalizations, and therefore constitute direct evidence of multiscale temporal processing of naturalistic stimuli by neurons in the AC. These neuronal subpopulations synchronize differently to local-field potentials, particularly in theta- and high frequency bands, and are informative to a different degree in terms of their spike rate. Interestingly, we also observed that both low and high frequency cortical oscillations can be highly informative about the listened calls. Our results suggest that multiscale neuronal processing allows for the precise and non-redundant representation of natural vocalizations in the AC.
Collapse
Affiliation(s)
- Francisco García-Rosales
- Institut für Zellbiologie und Neurowissenschaft, Goethe-Universität, 60438, Frankfurt/M., Germany.
| | - M Jerome Beetz
- Institut für Zellbiologie und Neurowissenschaft, Goethe-Universität, 60438, Frankfurt/M., Germany
- Department of Zoology II, University of Würzburg, Am Hubland, 97074, Würzburg, Germany
| | - Yuranny Cabral-Calderin
- MEG Labor, Brain Imaging Center, Goethe-Universität, 60528, Frankfurt/M., Germany
- German Resilience Center, University Medical Center Mainz, 55131, Mainz, Germany
| | - Manfred Kössl
- Institut für Zellbiologie und Neurowissenschaft, Goethe-Universität, 60438, Frankfurt/M., Germany
| | - Julio C Hechavarria
- Institut für Zellbiologie und Neurowissenschaft, Goethe-Universität, 60438, Frankfurt/M., Germany.
| |
Collapse
|
22
|
Steadman MA, Sumner CJ. Changes in Neuronal Representations of Consonants in the Ascending Auditory System and Their Role in Speech Recognition. Front Neurosci 2018; 12:671. [PMID: 30369863 PMCID: PMC6194309 DOI: 10.3389/fnins.2018.00671] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2018] [Accepted: 09/06/2018] [Indexed: 11/25/2022] Open
Abstract
A fundamental task of the ascending auditory system is to produce representations that facilitate the recognition of complex sounds. This is particularly challenging in the context of acoustic variability, such as that between different talkers producing the same phoneme. These representations are transformed as information is propagated throughout the ascending auditory system from the inner ear to the auditory cortex (AI). Investigating these transformations and their role in speech recognition is key to understanding hearing impairment and the development of future clinical interventions. Here, we obtained neural responses to an extensive set of natural vowel-consonant-vowel phoneme sequences, each produced by multiple talkers, in three stages of the auditory processing pathway. Auditory nerve (AN) representations were simulated using a model of the peripheral auditory system and extracellular neuronal activity was recorded in the inferior colliculus (IC) and primary auditory cortex (AI) of anaesthetized guinea pigs. A classifier was developed to examine the efficacy of these representations for recognizing the speech sounds. Individual neurons convey progressively less information from AN to AI. Nonetheless, at the population level, representations are sufficiently rich to facilitate recognition of consonants with a high degree of accuracy at all stages indicating a progression from a dense, redundant representation to a sparse, distributed one. We examined the timescale of the neural code for consonant recognition and found that optimal timescales increase throughout the ascending auditory system from a few milliseconds in the periphery to several tens of milliseconds in the cortex. Despite these longer timescales, we found little evidence to suggest that representations up to the level of AI become increasingly invariant to across-talker differences. Instead, our results support the idea that the role of the subcortical auditory system is one of dimensionality expansion, which could provide a basis for flexible classification of arbitrary speech sounds.
Collapse
Affiliation(s)
- Mark A. Steadman
- MRC Institute of Hearing Research, School of Medicine, The University of Nottingham, Nottingham, United Kingdom
- Department of Bioengineering, Imperial College London, London, United Kingdom
| | - Christian J. Sumner
- MRC Institute of Hearing Research, School of Medicine, The University of Nottingham, Nottingham, United Kingdom
| |
Collapse
|
23
|
Herz AV, Mathis A, Stemmler M. Periodic population codes: From a single circular variable to higher dimensions, multiple nested scales, and conceptual spaces. Curr Opin Neurobiol 2017; 46:99-108. [PMID: 28888183 DOI: 10.1016/j.conb.2017.07.005] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2017] [Revised: 07/06/2017] [Accepted: 07/19/2017] [Indexed: 12/27/2022]
Abstract
Across the nervous system, neurons often encode circular stimuli using tuning curves that are not sine or cosine functions, but that belong to the richer class of von Mises functions, which are periodic variants of Gaussians. For a population of neurons encoding a single circular variable with such canonical tuning curves, computing a simple population vector is the optimal read-out of the most likely stimulus. We argue that the advantages of population vector read-outs are so compelling that even the neural representation of the outside world's flat Euclidean geometry is curled up into a torus (a circle times a circle), creating the hexagonal activity patterns of mammalian grid cells. Here, the circular scale is not set a priori, so the nervous system can use multiple scales and gain fields to overcome the ambiguity inherent in periodic representations of linear variables. We review the experimental evidence for this framework and discuss its testable predictions and generalizations to more abstract grid-like neural representations.
Collapse
Affiliation(s)
- Andreas Vm Herz
- Bernstein Center for Computational Neuroscience Munich and Faculty of Biology, Ludwig-Maximilians-Universität München, Grosshadernerstrasse 2, 82152 Planegg-Martinsried, Germany.
| | - Alexander Mathis
- Department of Molecular and Cellular Biology and Center for Brain Science, Harvard University, 16 Divinity Avenue, Cambridge, MA 02138, USA; Werner Reichardt Centre for Integrative Neuroscience and Institute for Theoretical Physics, University of Tübingen, 72076 Tübingen, Germany
| | - Martin Stemmler
- Bernstein Center for Computational Neuroscience Munich and Faculty of Biology, Ludwig-Maximilians-Universität München, Grosshadernerstrasse 2, 82152 Planegg-Martinsried, Germany
| |
Collapse
|
24
|
Runyan CA, Piasini E, Panzeri S, Harvey CD. Distinct timescales of population coding across cortex. Nature 2017; 548:92-96. [PMID: 28723889 PMCID: PMC5859334 DOI: 10.1038/nature23020] [Citation(s) in RCA: 196] [Impact Index Per Article: 28.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2016] [Accepted: 06/13/2017] [Indexed: 01/21/2023]
Abstract
The cortex represents information across widely varying timescales1–5. For instance, sensory cortex encodes stimuli that fluctuate over few tens of milliseconds6,7, whereas in association cortex behavioral choices can require the maintenance of information over seconds8,9. However, it remains poorly understood if diverse timescales result mostly from features intrinsic to individual neurons or from neuronal population activity. This question is unanswered because the timescales of coding in populations of neurons have not been studied extensively, and population codes have not been compared systematically across cortical regions. Here we discovered that population codes can be essential to achieve long coding timescales. Furthermore, we found that the properties of population codes differ between sensory and association cortices. We compared coding for sensory stimuli and behavioral choices in auditory cortex (AC) and posterior parietal cortex (PPC) as mice performed a sound localization task. Auditory stimulus information was stronger in AC than in PPC, and both regions contained choice information. Although AC and PPC coded information by tiling in time neurons that were transiently informative for ~200 milliseconds, the areas had major differences in functional coupling between neurons, measured as activity correlations that could not be explained by task events. Coupling among PPC neurons was strong and extended over long time lags, whereas coupling among AC neurons was weak and short-lived. Stronger coupling in PPC led to a population code with long timescales and a representation of choice that remained consistent for approximately one second. In contrast, AC had a code with rapid fluctuations in stimulus and choice information over hundreds of milliseconds. Our results reveal that population codes differ across cortex and that coupling is a variable property of cortical populations that affects the timescale of information coding and the accuracy of behavior.
Collapse
Affiliation(s)
- Caroline A Runyan
- Department of Neurobiology, Harvard Medical School, Boston, Massachusetts 02115, USA
| | - Eugenio Piasini
- Neural Computation Laboratory, Istituto Italiano di Tecnologia, 38068 Rovereto, Italy
| | - Stefano Panzeri
- Neural Computation Laboratory, Istituto Italiano di Tecnologia, 38068 Rovereto, Italy
| | - Christopher D Harvey
- Department of Neurobiology, Harvard Medical School, Boston, Massachusetts 02115, USA
| |
Collapse
|
25
|
Atencio CA, Sharpee TO. Multidimensional receptive field processing by cat primary auditory cortical neurons. Neuroscience 2017; 359:130-141. [PMID: 28694174 DOI: 10.1016/j.neuroscience.2017.07.003] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2017] [Revised: 06/03/2017] [Accepted: 07/03/2017] [Indexed: 12/01/2022]
Abstract
The receptive fields of many auditory cortical neurons are multidimensional and are best represented by more than one stimulus feature. The number of these dimensions, their characteristics, and how they differ with stimulus context have been relatively unexplored. Standard methods that are often used to characterize multidimensional stimulus selectivity, such as spike-triggered covariance (STC) or maximally informative dimensions (MIDs), are either limited to Gaussian stimuli or are only able to recover a small number of stimulus features due to data limitations. An information theoretic extension of STC, the maximum noise entropy (MNE) model, can be used with non-Gaussian stimulus distributions to find an arbitrary number of stimulus dimensions. When we applied the MNE model to auditory cortical neurons, we often found more than two stimulus features that influenced neuronal firing. Excitatory and suppressive features coded different acoustic contexts: excitatory features encoded higher temporal and spectral modulations, while suppressive features had lower modulation frequency preferences. We found that the excitatory and suppressive features themselves were sensitive to stimulus context when we employed two stimuli that differed only in their short-term correlation structure: while the linear features were similar, the secondary features were strongly affected by stimulus statistics. These results show that multidimensional receptive field processing is influenced by feature type and stimulus context.
Collapse
Affiliation(s)
- Craig A Atencio
- Coleman Memorial Laboratory, UCSF Center for Integrative Neuroscience, Kavli Institute for Fundamental Neuroscience, Department of Otolaryngology-HNS, University of California, San Francisco, USA.
| | - Tatyana O Sharpee
- Computational Neurobiology Laboratory, The Salk Institute for Biological Studies, La Jolla, CA, USA; Center for Theoretical Biological Physics and Department of Physics, University of California, San Diego, La Jolla, CA, USA
| |
Collapse
|
26
|
Saha D, Sun W, Li C, Nizampatnam S, Padovano W, Chen Z, Chen A, Altan E, Lo R, Barbour DL, Raman B. Engaging and disengaging recurrent inhibition coincides with sensing and unsensing of a sensory stimulus. Nat Commun 2017; 8:15413. [PMID: 28534502 PMCID: PMC5457525 DOI: 10.1038/ncomms15413] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2016] [Accepted: 03/21/2017] [Indexed: 11/09/2022] Open
Abstract
Even simple sensory stimuli evoke neural responses that are dynamic and complex. Are the temporally patterned neural activities important for controlling the behavioral output? Here, we investigated this issue. Our results reveal that in the insect antennal lobe, due to circuit interactions, distinct neural ensembles are activated during and immediately following the termination of every odorant. Such non-overlapping response patterns are not observed even when the stimulus intensity or identities were changed. In addition, we find that ON and OFF ensemble neural activities differ in their ability to recruit recurrent inhibition, entrain field-potential oscillations and more importantly in their relevance to behaviour (initiate versus reset conditioned responses). Notably, we find that a strikingly similar strategy is also used for encoding sound onsets and offsets in the marmoset auditory cortex. In sum, our results suggest a general approach where recurrent inhibition is associated with stimulus ‘recognition' and ‘derecognition'. Sensory stimuli evoke temporally dynamic responses. Here the authors report that responses to odour onset and offset are orthogonally represented in the locust antennal lobe, differentially entrain oscillations, and propose a model in which they are necessary for initiation and termination of behaviour.
Collapse
Affiliation(s)
- Debajit Saha
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Wensheng Sun
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Chao Li
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Srinath Nizampatnam
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - William Padovano
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Zhengdao Chen
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Alex Chen
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Ege Altan
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Ray Lo
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Dennis L Barbour
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| | - Baranidharan Raman
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri 63130, USA
| |
Collapse
|
27
|
Huang N, Elhilali M. Auditory salience using natural soundscapes. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:2163. [PMID: 28372080 PMCID: PMC6909985 DOI: 10.1121/1.4979055] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/07/2016] [Revised: 03/09/2017] [Accepted: 03/10/2017] [Indexed: 05/26/2023]
Abstract
Salience describes the phenomenon by which an object stands out from a scene. While its underlying processes are extensively studied in vision, mechanisms of auditory salience remain largely unknown. Previous studies have used well-controlled auditory scenes to shed light on some of the acoustic attributes that drive the salience of sound events. Unfortunately, the use of constrained stimuli in addition to a lack of well-established benchmarks of salience judgments hampers the development of comprehensive theories of sensory-driven auditory attention. The present study explores auditory salience in a set of dynamic natural scenes. A behavioral measure of salience is collected by having human volunteers listen to two concurrent scenes and indicate continuously which one attracts their attention. By using natural scenes, the study takes a data-driven rather than experimenter-driven approach to exploring the parameters of auditory salience. The findings indicate that the space of auditory salience is multidimensional (spanning loudness, pitch, spectral shape, as well as other acoustic attributes), nonlinear and highly context-dependent. Importantly, the results indicate that contextual information about the entire scene over both short and long scales needs to be considered in order to properly account for perceptual judgments of salience.
Collapse
Affiliation(s)
- Nicholas Huang
- Department of Biomedical Engineering, The Johns Hopkins University, Baltimore, Maryland 21218, USA
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Center for Language and Speech Processing, The Johns Hopkins University, Baltimore, Maryland 21218, USA
| |
Collapse
|
28
|
Bender DA, Ni R, Barbour DL. Spontaneous activity is correlated with coding density in primary auditory cortex. J Neurophysiol 2016; 116:2789-2798. [PMID: 27707812 DOI: 10.1152/jn.00474.2016] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2016] [Accepted: 09/30/2016] [Indexed: 11/22/2022] Open
Abstract
Sensory neurons across sensory modalities and specific processing areas have diverse levels of spontaneous firing rates (SFRs) in the absence of sensory stimuli. However, the functional significance of this spontaneous activity is not well-understood. Previous studies in the auditory system have demonstrated that different levels of spontaneous activity are correlated with a variety of physiological and anatomic properties, suggesting that neurons with differing SFRs make unique contributions to the encoding of auditory stimuli. Additionally, altered SFRs are a correlate of tinnitus, arising in several auditory areas after exposure to ototoxic substances and noise trauma. In this study, we recorded single-unit activity from primary auditory cortex of awake marmoset monkeys while delivering wide-band random-spectrum stimuli and white Gaussian noise (WGN) to examine any divergences in stimulus encoding properties across SFR classes. We found that higher levels of spontaneous activity were associated with both higher levels of activation relative to suppression across a variety of wide-band stimuli and higher driven rates in response to WGN. Moreover, response latencies to WGN were negatively correlated with the level of activation in response to both stimulus types. These findings are consistent with a novel view of the role spontaneous spiking may play during normal stimulus processing in primary auditory cortex and how it may malfunction in cases of tinnitus.
Collapse
Affiliation(s)
- David A Bender
- Laboratory of Sensory Neuroscience and Neuroengineering, Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri; and.,Department of Biology, Washington University in St. Louis, St. Louis, Missouri
| | - Ruiye Ni
- Laboratory of Sensory Neuroscience and Neuroengineering, Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri; and
| | - Dennis L Barbour
- Laboratory of Sensory Neuroscience and Neuroengineering, Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, Missouri; and
| |
Collapse
|
29
|
Lee CM, Osman AF, Volgushev M, Escabí MA, Read HL. Neural spike-timing patterns vary with sound shape and periodicity in three auditory cortical fields. J Neurophysiol 2016; 115:1886-904. [PMID: 26843599 DOI: 10.1152/jn.00784.2015] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2015] [Accepted: 01/29/2016] [Indexed: 11/22/2022] Open
Abstract
Mammals perceive a wide range of temporal cues in natural sounds, and the auditory cortex is essential for their detection and discrimination. The rat primary (A1), ventral (VAF), and caudal suprarhinal (cSRAF) auditory cortical fields have separate thalamocortical pathways that may support unique temporal cue sensitivities. To explore this, we record responses of single neurons in the three fields to variations in envelope shape and modulation frequency of periodic noise sequences. Spike rate, relative synchrony, and first-spike latency metrics have previously been used to quantify neural sensitivities to temporal sound cues; however, such metrics do not measure absolute spike timing of sustained responses to sound shape. To address this, in this study we quantify two forms of spike-timing precision, jitter, and reliability. In all three fields, we find that jitter decreases logarithmically with increase in the basis spline (B-spline) cutoff frequency used to shape the sound envelope. In contrast, reliability decreases logarithmically with increase in sound envelope modulation frequency. In A1, jitter and reliability vary independently, whereas in ventral cortical fields, jitter and reliability covary. Jitter time scales increase (A1 < VAF < cSRAF) and modulation frequency upper cutoffs decrease (A1 > VAF > cSRAF) with ventral progression from A1. These results suggest a transition from independent encoding of shape and periodicity sound cues on short time scales in A1 to a joint encoding of these same cues on longer time scales in ventral nonprimary cortices.
Collapse
Affiliation(s)
- Christopher M Lee
- Department of Psychology, University of Connecticut, Storrs, Connecticut
| | - Ahmad F Osman
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and
| | - Maxim Volgushev
- Department of Psychology, University of Connecticut, Storrs, Connecticut
| | - Monty A Escabí
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut
| | - Heather L Read
- Department of Psychology, University of Connecticut, Storrs, Connecticut; Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut; and
| |
Collapse
|
30
|
Blackwell JM, Taillefumier TO, Natan RG, Carruthers IM, Magnasco MO, Geffen MN. Stable encoding of sounds over a broad range of statistical parameters in the auditory cortex. Eur J Neurosci 2016; 43:751-64. [PMID: 26663571 PMCID: PMC5021175 DOI: 10.1111/ejn.13144] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2015] [Revised: 11/22/2015] [Accepted: 12/01/2015] [Indexed: 11/29/2022]
Abstract
Natural auditory scenes possess highly structured statistical regularities, which are dictated by the physics of sound production in nature, such as scale‐invariance. We recently identified that natural water sounds exhibit a particular type of scale invariance, in which the temporal modulation within spectral bands scales with the centre frequency of the band. Here, we tested how neurons in the mammalian primary auditory cortex encode sounds that exhibit this property, but differ in their statistical parameters. The stimuli varied in spectro‐temporal density and cyclo‐temporal statistics over several orders of magnitude, corresponding to a range of water‐like percepts, from pattering of rain to a slow stream. We recorded neuronal activity in the primary auditory cortex of awake rats presented with these stimuli. The responses of the majority of individual neurons were selective for a subset of stimuli with specific statistics. However, as a neuronal population, the responses were remarkably stable over large changes in stimulus statistics, exhibiting a similar range in firing rate, response strength, variability and information rate, and only minor variation in receptive field parameters. This pattern of neuronal responses suggests a potentially general principle for cortical encoding of complex acoustic scenes: while individual cortical neurons exhibit selectivity for specific statistical features, a neuronal population preserves a constant response structure across a broad range of statistical parameters.
Collapse
Affiliation(s)
- Jennifer M Blackwell
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Thibaud O Taillefumier
- Center for Physics and Biology, Rockefeller University, New York, NY, USA.,Lewis-Sigler Institute for Integrative Genomics, Princeton University, Princeton, NJ, USA
| | - Ryan G Natan
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Isaac M Carruthers
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Marcelo O Magnasco
- Center for Physics and Biology, Rockefeller University, New York, NY, USA
| | - Maria N Geffen
- Department of Otorhinolaryngology and Head and Neck Surgery, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA.,Center for Physics and Biology, Rockefeller University, New York, NY, USA
| |
Collapse
|
31
|
Chambers JD, Burkitt AN, Grayden DB. Computational neural modelling of auditory cortical receptive fields. BMC Neurosci 2015. [PMCID: PMC4697619 DOI: 10.1186/1471-2202-16-s1-p85] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
|
32
|
Carlin MA, Elhilali M. A Framework for Speech Activity Detection Using Adaptive Auditory Receptive Fields. IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING 2015; 23:2422-2433. [PMID: 29904642 PMCID: PMC5997283 DOI: 10.1109/taslp.2015.2481179] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
One of the hallmarks of sound processing in the brain is the ability of the nervous system to adapt to changing behavioral demands and surrounding soundscapes. It can dynamically shift sensory and cognitive resources to focus on relevant sounds. Neurophysiological studies indicate that this ability is supported by adaptively retuning the shapes of cortical spectro-temporal receptive fields (STRFs) to enhance features of target sounds while suppressing those of task-irrelevant distractors. Because an important component of human communication is the ability of a listener to dynamically track speech in noisy environments, the solution obtained by auditory neurophysiology implies a useful adaptation strategy for speech activity detection (SAD). SAD is an important first step in a number of automated speech processing systems, and performance is often reduced in highly noisy environments. In this paper, we describe how task-driven adaptation is induced in an ensemble of neurophysiological STRFs, and show how speech-adapted STRFs reorient themselves to enhance spectro-temporal modulations of speech while suppressing those associated with a variety of nonspeech sounds. We then show how an adapted ensemble of STRFs can better detect speech in unseen noisy environments compared to an unadapted ensemble and a noise-robust baseline. Finally, we use a stimulus reconstruction task to demonstrate how the adapted STRF ensemble better captures the spectrotemporal modulations of attended speech in clean and noisy conditions. Our results suggest that a biologically plausible adaptation framework can be applied to speech processing systems to dynamically adapt feature representations for improving noise robustness.
Collapse
Affiliation(s)
- Michael A Carlin
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218 USA
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218 USA
| |
Collapse
|
33
|
|
34
|
Schafer PB, Jin DZ. Noise-Robust Speech Recognition Through Auditory Feature Detection and Spike Sequence Decoding. Neural Comput 2014; 26:523-56. [DOI: 10.1162/neco_a_00557] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Speech recognition in noisy conditions is a major challenge for computer systems, but the human brain performs it routinely and accurately. Automatic speech recognition (ASR) systems that are inspired by neuroscience can potentially bridge the performance gap between humans and machines. We present a system for noise-robust isolated word recognition that works by decoding sequences of spikes from a population of simulated auditory feature-detecting neurons. Each neuron is trained to respond selectively to a brief spectrotemporal pattern, or feature, drawn from the simulated auditory nerve response to speech. The neural population conveys the time-dependent structure of a sound by its sequence of spikes. We compare two methods for decoding the spike sequences—one using a hidden Markov model–based recognizer, the other using a novel template-based recognition scheme. In the latter case, words are recognized by comparing their spike sequences to template sequences obtained from clean training data, using a similarity measure based on the length of the longest common sub-sequence. Using isolated spoken digits from the AURORA-2 database, we show that our combined system outperforms a state-of-the-art robust speech recognizer at low signal-to-noise ratios. Both the spike-based encoding scheme and the template-based decoding offer gains in noise robustness over traditional speech recognition methods. Our system highlights potential advantages of spike-based acoustic coding and provides a biologically motivated framework for robust ASR development.
Collapse
Affiliation(s)
- Phillip B. Schafer
- Department of Physics and Center for Neural Engineering, The Pennsylvania State University, University Park, PA 16802, U.S.A
| | - Dezhe Z. Jin
- Department of Physics and Center for Neural Engineering, The Pennsylvania State University, University Park, PA 16802, U.S.A
| |
Collapse
|
35
|
Micheyl C, Schrater PR, Oxenham AJ. Auditory frequency and intensity discrimination explained using a cortical population rate code. PLoS Comput Biol 2013; 9:e1003336. [PMID: 24244142 PMCID: PMC3828126 DOI: 10.1371/journal.pcbi.1003336] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2013] [Accepted: 09/27/2013] [Indexed: 11/18/2022] Open
Abstract
The nature of the neural codes for pitch and loudness, two basic auditory attributes, has been a key question in neuroscience for over century. A currently widespread view is that sound intensity (subjectively, loudness) is encoded in spike rates, whereas sound frequency (subjectively, pitch) is encoded in precise spike timing. Here, using information-theoretic analyses, we show that the spike rates of a population of virtual neural units with frequency-tuning and spike-count correlation characteristics similar to those measured in the primary auditory cortex of primates, contain sufficient statistical information to account for the smallest frequency-discrimination thresholds measured in human listeners. The same population, and the same spike-rate code, can also account for the intensity-discrimination thresholds of humans. These results demonstrate the viability of a unified rate-based cortical population code for both sound frequency (pitch) and sound intensity (loudness), and thus suggest a resolution to a long-standing puzzle in auditory neuroscience.
Collapse
Affiliation(s)
- Christophe Micheyl
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
- * E-mail:
| | - Paul R. Schrater
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
- Department of Computer Science, University of Minnesota, Minneapolis, Minnesota, United States of America
| | - Andrew J. Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota, United States of America
- Department of Otolaryngology, University of Minnesota, Minneapolis, Minnesota, United States of America
| |
Collapse
|
36
|
Cortical inhibition reduces information redundancy at presentation of communication sounds in the primary auditory cortex. J Neurosci 2013; 33:10713-28. [PMID: 23804094 DOI: 10.1523/jneurosci.0079-13.2013] [Citation(s) in RCA: 36] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
In all sensory modalities, intracortical inhibition shapes the functional properties of cortical neurons but also influences the responses to natural stimuli. Studies performed in various species have revealed that auditory cortex neurons respond to conspecific vocalizations by temporal spike patterns displaying a high trial-to-trial reliability, which might result from precise timing between excitation and inhibition. Studying the guinea pig auditory cortex, we show that partial blockage of GABAA receptors by gabazine (GBZ) application (10 μm, a concentration that promotes expansion of cortical receptive fields) increased the evoked firing rate and the spike-timing reliability during presentation of communication sounds (conspecific and heterospecific vocalizations), whereas GABAB receptor antagonists [10 μm saclofen; 10-50 μm CGP55845 (p-3-aminopropyl-p-diethoxymethyl phosphoric acid)] had nonsignificant effects. Computing mutual information (MI) from the responses to vocalizations using either the evoked firing rate or the temporal spike patterns revealed that GBZ application increased the MI derived from the activity of single cortical site but did not change the MI derived from population activity. In addition, quantification of information redundancy showed that GBZ significantly increased redundancy at the population level. This result suggests that a potential role of intracortical inhibition is to reduce information redundancy during the processing of natural stimuli.
Collapse
|
37
|
Pre-attentive, context-specific representation of fear memory in the auditory cortex of rat. PLoS One 2013; 8:e63655. [PMID: 23671691 PMCID: PMC3646040 DOI: 10.1371/journal.pone.0063655] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2012] [Accepted: 04/04/2013] [Indexed: 11/29/2022] Open
Abstract
Neural representation in the auditory cortex is rapidly modulated by both top-down attention and bottom-up stimulus properties, in order to improve perception in a given context. Learning-induced, pre-attentive, map plasticity has been also studied in the anesthetized cortex; however, little attention has been paid to rapid, context-dependent modulation. We hypothesize that context-specific learning leads to pre-attentively modulated, multiplex representation in the auditory cortex. Here, we investigate map plasticity in the auditory cortices of anesthetized rats conditioned in a context-dependent manner, such that a conditioned stimulus (CS) of a 20-kHz tone and an unconditioned stimulus (US) of a mild electrical shock were associated only under a noisy auditory context, but not in silence. After the conditioning, although no distinct plasticity was found in the tonotopic map, tone-evoked responses were more noise-resistive than pre-conditioning. Yet, the conditioned group showed a reduced spread of activation to each tone with noise, but not with silence, associated with a sharpening of frequency tuning. The encoding accuracy index of neurons showed that conditioning deteriorated the accuracy of tone-frequency representations in noisy condition at off-CS regions, but not at CS regions, suggesting that arbitrary tones around the frequency of the CS were more likely perceived as the CS in a specific context, where CS was associated with US. These results together demonstrate that learning-induced plasticity in the auditory cortex occurs in a context-dependent manner.
Collapse
|
38
|
Zheng Y, Escabí MA. Proportional spike-timing precision and firing reliability underlie efficient temporal processing of periodicity and envelope shape cues. J Neurophysiol 2013; 110:587-606. [PMID: 23636724 DOI: 10.1152/jn.01080.2010] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Temporal sound cues are essential for sound recognition, pitch, rhythm, and timbre perception, yet how auditory neurons encode such cues is subject of ongoing debate. Rate coding theories propose that temporal sound features are represented by rate tuned modulation filters. However, overwhelming evidence also suggests that precise spike timing is an essential attribute of the neural code. Here we demonstrate that single neurons in the auditory midbrain employ a proportional code in which spike-timing precision and firing reliability covary with the sound envelope cues to provide an efficient representation of the stimulus. Spike-timing precision varied systematically with the timescale and shape of the sound envelope and yet was largely independent of the sound modulation frequency, a prominent cue for pitch. In contrast, spike-count reliability was strongly affected by the modulation frequency. Spike-timing precision extends from sub-millisecond for brief transient sounds up to tens of milliseconds for sounds with slow-varying envelope. Information theoretic analysis further confirms that spike-timing precision depends strongly on the sound envelope shape, while firing reliability was strongly affected by the sound modulation frequency. Both the information efficiency and total information were limited by the firing reliability and spike-timing precision in a manner that reflected the sound structure. This result supports a temporal coding strategy in the auditory midbrain where proportional changes in spike-timing precision and firing reliability can efficiently signal shape and periodicity temporal cues.
Collapse
Affiliation(s)
- Y Zheng
- Department of Biomedical Engineering, University of Connecticut, Storrs, CT 06269, USA
| | | |
Collapse
|
39
|
Zion Golumbic EM, Ding N, Bickel S, Lakatos P, Schevon CA, McKhann GM, Goodman RR, Emerson R, Mehta AD, Simon JZ, Poeppel D, Schroeder CE. Mechanisms underlying selective neuronal tracking of attended speech at a "cocktail party". Neuron 2013; 77:980-91. [PMID: 23473326 DOI: 10.1016/j.neuron.2012.12.037] [Citation(s) in RCA: 518] [Impact Index Per Article: 47.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/21/2012] [Indexed: 11/26/2022]
Abstract
The ability to focus on and understand one talker in a noisy social environment is a critical social-cognitive capacity, whose underlying neuronal mechanisms are unclear. We investigated the manner in which speech streams are represented in brain activity and the way that selective attention governs the brain's representation of speech using a "Cocktail Party" paradigm, coupled with direct recordings from the cortical surface in surgical epilepsy patients. We find that brain activity dynamically tracks speech streams using both low-frequency phase and high-frequency amplitude fluctuations and that optimal encoding likely combines the two. In and near low-level auditory cortices, attention "modulates" the representation by enhancing cortical tracking of attended speech streams, but ignored speech remains represented. In higher-order regions, the representation appears to become more "selective," in that there is no detectable tracking of ignored speech. This selectivity itself seems to sharpen as a sentence unfolds.
Collapse
Affiliation(s)
- Elana M Zion Golumbic
- Department of Psychiatry, Columbia University College of Physicians and Surgeons, New York, NY, USA
| | | | | | | | | | | | | | | | | | | | | | | |
Collapse
|
40
|
Gaucher Q, Huetz C, Gourévitch B, Laudanski J, Occelli F, Edeline JM. How do auditory cortex neurons represent communication sounds? Hear Res 2013; 305:102-12. [PMID: 23603138 DOI: 10.1016/j.heares.2013.03.011] [Citation(s) in RCA: 33] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/03/2012] [Revised: 03/18/2013] [Accepted: 03/26/2013] [Indexed: 11/30/2022]
Abstract
A major goal in auditory neuroscience is to characterize how communication sounds are represented at the cortical level. The present review aims at investigating the role of auditory cortex in the processing of speech, bird songs and other vocalizations, which all are spectrally and temporally highly structured sounds. Whereas earlier studies have simply looked for neurons exhibiting higher firing rates to particular conspecific vocalizations over their modified, artificially synthesized versions, more recent studies determined the coding capacity of temporal spike patterns, which are prominent in primary and non-primary areas (and also in non-auditory cortical areas). In several cases, this information seems to be correlated with the behavioral performance of human or animal subjects, suggesting that spike-timing based coding strategies might set the foundations of our perceptive abilities. Also, it is now clear that the responses of auditory cortex neurons are highly nonlinear and that their responses to natural stimuli cannot be predicted from their responses to artificial stimuli such as moving ripples and broadband noises. Since auditory cortex neurons cannot follow rapid fluctuations of the vocalizations envelope, they only respond at specific time points during communication sounds, which can serve as temporal markers for integrating the temporal and spectral processing taking place at subcortical relays. Thus, the temporal sparse code of auditory cortex neurons can be considered as a first step for generating high level representations of communication sounds independent of the acoustic characteristic of these sounds. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives".
Collapse
Affiliation(s)
- Quentin Gaucher
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Université Paris-Sud, Bâtiment 446, 91405 Orsay cedex, France
| | | | | | | | | | | |
Collapse
|
41
|
Carlin MA, Elhilali M. Sustained firing of model central auditory neurons yields a discriminative spectro-temporal representation for natural sounds. PLoS Comput Biol 2013; 9:e1002982. [PMID: 23555217 PMCID: PMC3610626 DOI: 10.1371/journal.pcbi.1002982] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2012] [Accepted: 01/25/2013] [Indexed: 11/19/2022] Open
Abstract
The processing characteristics of neurons in the central auditory system are directly shaped by and reflect the statistics of natural acoustic environments, but the principles that govern the relationship between natural sound ensembles and observed responses in neurophysiological studies remain unclear. In particular, accumulating evidence suggests the presence of a code based on sustained neural firing rates, where central auditory neurons exhibit strong, persistent responses to their preferred stimuli. Such a strategy can indicate the presence of ongoing sounds, is involved in parsing complex auditory scenes, and may play a role in matching neural dynamics to varying time scales in acoustic signals. In this paper, we describe a computational framework for exploring the influence of a code based on sustained firing rates on the shape of the spectro-temporal receptive field (STRF), a linear kernel that maps a spectro-temporal acoustic stimulus to the instantaneous firing rate of a central auditory neuron. We demonstrate the emergence of richly structured STRFs that capture the structure of natural sounds over a wide range of timescales, and show how the emergent ensembles resemble those commonly reported in physiological studies. Furthermore, we compare ensembles that optimize a sustained firing code with one that optimizes a sparse code, another widely considered coding strategy, and suggest how the resulting population responses are not mutually exclusive. Finally, we demonstrate how the emergent ensembles contour the high-energy spectro-temporal modulations of natural sounds, forming a discriminative representation that captures the full range of modulation statistics that characterize natural sound ensembles. These findings have direct implications for our understanding of how sensory systems encode the informative components of natural stimuli and potentially facilitate multi-sensory integration.
Collapse
Affiliation(s)
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, The Center for Language and Speech Processing, Johns Hopkins University, Baltimore, Maryland, United States of America
| |
Collapse
|
42
|
A phase transition in the first passage of a Brownian process through a fluctuating boundary with implications for neural coding. Proc Natl Acad Sci U S A 2013; 110:E1438-43. [PMID: 23536302 DOI: 10.1073/pnas.1212479110] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Finding the first time a fluctuating quantity reaches a given boundary is a deceptively simple-looking problem of vast practical importance in physics, biology, chemistry, neuroscience, economics, and industrial engineering. Problems in which the bound to be traversed is itself a fluctuating function of time include widely studied problems in neural coding, such as neuronal integrators with irregular inputs and internal noise. We show that the probability p(t) that a Gauss-Markov process will first exceed the boundary at time t suffers a phase transition as a function of the roughness of the boundary, as measured by its Hölder exponent H. The critical value occurs when the roughness of the boundary equals the roughness of the process, so for diffusive processes the critical value is Hc = 1/2. For smoother boundaries, H > 1/2, the probability density is a continuous function of time. For rougher boundaries, H < 1/2, the probability is concentrated on a Cantor-like set of zero measure: the probability density becomes divergent, almost everywhere either zero or infinity. The critical point Hc = 1/2 corresponds to a widely studied case in the theory of neural coding, in which the external input integrated by a model neuron is a white-noise process, as in the case of uncorrelated but precisely balanced excitatory and inhibitory inputs. We argue that this transition corresponds to a sharp boundary between rate codes, in which the neural firing probability varies smoothly, and temporal codes, in which the neuron fires at sharply defined times regardless of the intensity of internal noise.
Collapse
|
43
|
Neurogenetics and auditory processing in developmental dyslexia. Curr Opin Neurobiol 2013; 23:37-42. [PMID: 23040541 DOI: 10.1016/j.conb.2012.09.003] [Citation(s) in RCA: 72] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2012] [Revised: 09/06/2012] [Accepted: 09/16/2012] [Indexed: 01/18/2023]
|
44
|
Oppenheim JN, Magnasco MO. Human time-frequency acuity beats the Fourier uncertainty principle. PHYSICAL REVIEW LETTERS 2013; 110:044301. [PMID: 25166166 DOI: 10.1103/physrevlett.110.044301] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/24/2012] [Revised: 11/13/2012] [Indexed: 06/03/2023]
Abstract
The time-frequency uncertainty principle states that the product of the temporal and frequency extents of a signal cannot be smaller than 1/(4 π). We study human ability to simultaneously judge the frequency and the timing of a sound. Our subjects often exceeded the uncertainty limit, sometimes by more than tenfold, mostly through remarkable timing acuity. Our results establish a lower bound for the nonlinearity and complexity of the algorithms employed by our brains in parsing transient sounds, rule out simple "linear filter" models of early auditory processing, and highlight timing acuity as a central feature in auditory object processing.
Collapse
Affiliation(s)
- Jacob N Oppenheim
- Laboratory of Mathematical Physics, Rockefeller University, New York, New York 10065, USA
| | - Marcelo O Magnasco
- Laboratory of Mathematical Physics, Rockefeller University, New York, New York 10065, USA
| |
Collapse
|
45
|
Laudanski J, Edeline JM, Huetz C. Differences between spectro-temporal receptive fields derived from artificial and natural stimuli in the auditory cortex. PLoS One 2012; 7:e50539. [PMID: 23209771 PMCID: PMC3507792 DOI: 10.1371/journal.pone.0050539] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2012] [Accepted: 10/25/2012] [Indexed: 11/25/2022] Open
Abstract
Spectro-temporal properties of auditory cortex neurons have been extensively studied with artificial sounds but it is still unclear whether they help in understanding neuronal responses to communication sounds. Here, we directly compared spectro-temporal receptive fields (STRFs) obtained from the same neurons using both artificial stimuli (dynamic moving ripples, DMRs) and natural stimuli (conspecific vocalizations) that were matched in terms of spectral content, average power and modulation spectrum. On a population of auditory cortex neurons exhibiting reliable tuning curves when tested with pure tones, significant STRFs were obtained for 62% of the cells with vocalizations and 68% with DMR. However, for many cells with significant vocalization-derived STRFs (STRFvoc) and DMR-derived STRFs (STRFdmr), the BF, latency, bandwidth and global STRFs shape differed more than what would be predicted by spiking responses simulated by a linear model based on a non-homogenous Poisson process. Moreover STRFvoc predicted neural responses to vocalizations more accurately than STRFdmr predicted neural response to DMRs, despite similar spike-timing reliability for both sets of stimuli. Cortical bursts, which potentially introduce nonlinearities in evoked responses, did not explain the differences between STRFvoc and STRFdmr. Altogether, these results suggest that the nonlinearity of auditory cortical responses makes it difficult to predict responses to communication sounds from STRFs computed from artificial stimuli.
Collapse
Affiliation(s)
- Jonathan Laudanski
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| | - Jean-Marc Edeline
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
- * E-mail:
| | - Chloé Huetz
- Centre de Neurosciences Paris-Sud (CNPS), CNRS UMR 8195, Orsay, France
- Centre de Neurosciences Paris-Sud, Université Paris-Sud, Orsay, France
| |
Collapse
|
46
|
Rabang CF, Parthasarathy A, Venkataraman Y, Fisher ZL, Gardner SM, Bartlett EL. A computational model of inferior colliculus responses to amplitude modulated sounds in young and aged rats. Front Neural Circuits 2012; 6:77. [PMID: 23129994 PMCID: PMC3487458 DOI: 10.3389/fncir.2012.00077] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2012] [Accepted: 10/05/2012] [Indexed: 12/03/2022] Open
Abstract
The inferior colliculus (IC) receives ascending excitatory and inhibitory inputs from multiple sources, but how these auditory inputs converge to generate IC spike patterns is poorly understood. Simulating patterns of in vivo spike train data from cellular and synaptic models creates a powerful framework to identify factors that contribute to changes in IC responses, such as those resulting in age-related loss of temporal processing. A conductance-based single neuron IC model was constructed, and its responses were compared to those observed during in vivo IC recordings in rats. IC spike patterns were evoked using amplitude-modulated tone or noise carriers at 20–40 dB above threshold and were classified as low-pass, band-pass, band-reject, all-pass, or complex based on their rate modulation transfer function tuning shape. Their temporal modulation transfer functions were also measured. These spike patterns provided experimental measures of rate, vector strength, and firing pattern for comparison with model outputs. Patterns of excitatory and inhibitory synaptic convergence to IC neurons were based on anatomical studies and generalized input tuning for modulation frequency. Responses of modeled ascending inputs were derived from experimental data from previous studies. Adapting and sustained IC intrinsic models were created, with adaptation created via calcium-activated potassium currents. Short-term synaptic plasticity was incorporated into the model in the form of synaptic depression, which was shown to have a substantial effect on the magnitude and time course of the IC response. The most commonly observed IC response sub-types were recreated and enabled dissociation of inherited response properties from those that were generated in IC. Furthermore, the model was used to make predictions about the consequences of reduction in inhibition for age-related loss of temporal processing due to a reduction in GABA seen anatomically with age.
Collapse
Affiliation(s)
- Cal F Rabang
- Weldon School of Biomedical Engineering, Purdue University West Lafayette, IN, USA
| | | | | | | | | | | |
Collapse
|
47
|
Patil K, Pressnitzer D, Shamma S, Elhilali M. Music in our ears: the biological bases of musical timbre perception. PLoS Comput Biol 2012; 8:e1002759. [PMID: 23133363 PMCID: PMC3486808 DOI: 10.1371/journal.pcbi.1002759] [Citation(s) in RCA: 50] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2012] [Accepted: 09/12/2012] [Indexed: 11/18/2022] Open
Abstract
Timbre is the attribute of sound that allows humans and other animals to distinguish among different sound sources. Studies based on psychophysical judgments of musical timbre, ecological analyses of sound's physical characteristics as well as machine learning approaches have all suggested that timbre is a multifaceted attribute that invokes both spectral and temporal sound features. Here, we explored the neural underpinnings of musical timbre. We used a neuro-computational framework based on spectro-temporal receptive fields, recorded from over a thousand neurons in the mammalian primary auditory cortex as well as from simulated cortical neurons, augmented with a nonlinear classifier. The model was able to perform robust instrument classification irrespective of pitch and playing style, with an accuracy of 98.7%. Using the same front end, the model was also able to reproduce perceptual distance judgments between timbres as perceived by human listeners. The study demonstrates that joint spectro-temporal features, such as those observed in the mammalian primary auditory cortex, are critical to provide the rich-enough representation necessary to account for perceptual judgments of timbre by human listeners, as well as recognition of musical instruments.
Collapse
Affiliation(s)
- Kailash Patil
- Department of Electrical and Computer Engineering, Center for Language and Speech Processing, Johns Hopkins University, Baltimore, Maryland, United States of America
| | - Daniel Pressnitzer
- Laboratoire Psychologie de la Perception, CNRS-Université Paris Descartes & DEC, Ecole normale supérieure, Paris, France
| | - Shihab Shamma
- Department of Electrical and Computer Engineering and Institute for Systems Research, University of Maryland, College Park, Maryland, United States of America
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Center for Language and Speech Processing, Johns Hopkins University, Baltimore, Maryland, United States of America
- * E-mail:
| |
Collapse
|
48
|
Abstract
In many sensory systems, the latency of spike responses of individual neurons is found to be tuned for stimulus features and proposed to be used as a coding strategy. Whether the spike latency tuning is simply relayed along sensory ascending pathways or generated by local circuits remains unclear. Here, in vivo whole-cell recordings from rat auditory cortical neurons in layer 4 revealed that the onset latency of their aggregate thalamic input exhibited nearly flat tuning for sound frequency, whereas their spike latency tuning was much sharper with a broadly expanded dynamic range. This suggests that the spike latency tuning is not simply inherited from the thalamus, but can be largely reconstructed by local circuits in the cortex. Dissecting of thalamocortical circuits and neural modeling further revealed that broadly tuned intracortical inhibition prolongs the integration time for spike generation preferentially at off-optimal frequencies, while sharply tuned intracortical excitation shortens it selectively at the optimal frequency. Such push and pull mechanisms mediated likely by feedforward excitatory and inhibitory inputs respectively greatly sharpen the spike latency tuning and expand its dynamic range. The modulation of integration time by thalamocortical-like circuits may represent an efficient strategy for converting information spatially coded in synaptic strength to temporal representation.
Collapse
|
49
|
Zion Golumbic EM, Poeppel D, Schroeder CE. Temporal context in speech processing and attentional stream selection: a behavioral and neural perspective. BRAIN AND LANGUAGE 2012; 122:151-61. [PMID: 22285024 PMCID: PMC3340429 DOI: 10.1016/j.bandl.2011.12.010] [Citation(s) in RCA: 108] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/07/2011] [Revised: 12/14/2011] [Accepted: 12/16/2011] [Indexed: 05/04/2023]
Abstract
The human capacity for processing speech is remarkable, especially given that information in speech unfolds over multiple time scales concurrently. Similarly notable is our ability to filter out of extraneous sounds and focus our attention on one conversation, epitomized by the 'Cocktail Party' effect. Yet, the neural mechanisms underlying on-line speech decoding and attentional stream selection are not well understood. We review findings from behavioral and neurophysiological investigations that underscore the importance of the temporal structure of speech for achieving these perceptual feats. We discuss the hypothesis that entrainment of ambient neuronal oscillations to speech's temporal structure, across multiple time-scales, serves to facilitate its decoding and underlies the selection of an attended speech stream over other competing input. In this regard, speech decoding and attentional stream selection are examples of 'Active Sensing', emphasizing an interaction between proactive and predictive top-down modulation of neuronal dynamics and bottom-up sensory input.
Collapse
Affiliation(s)
- Elana M Zion Golumbic
- Department of Psychiatry, Columbia University Medical Center, 710 W 168th St., New York, NY 10032, USA.
| | | | | |
Collapse
|
50
|
Precise feature based time scales and frequency decorrelation lead to a sparse auditory code. J Neurosci 2012; 32:8454-68. [PMID: 22723685 DOI: 10.1523/jneurosci.6506-11.2012] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Sparse redundancy reducing codes have been proposed as efficient strategies for representing sensory stimuli. A prevailing hypothesis suggests that sensory representations shift from dense redundant codes in the periphery to selective sparse codes in cortex. We propose an alternative framework where sparseness and redundancy depend on sensory integration time scales and demonstrate that the central nucleus of the inferior colliculus (ICC) of cats encodes sound features by precise sparse spike trains. Direct comparisons with auditory cortical neurons demonstrate that ICC responses were sparse and uncorrelated as long as the spike train time scales were matched to the sensory integration time scales relevant to ICC neurons. Intriguingly, correlated spiking in the ICC was substantially lower than predicted by linear or nonlinear models and strictly observed for neurons with best frequencies within a "critical band," the hallmark of perceptual frequency resolution in mammals. This is consistent with a sparse asynchronous code throughout much of the ICC and a complementary correlation code within a critical band that may allow grouping of perceptually relevant cues.
Collapse
|