1
|
Gorina-Careta N, Arenillas-Alcón S, Puertollano M, Mondéjar-Segovia A, Ijjou-Kadiri S, Costa-Faidella J, Gómez-Roig MD, Escera C. Exposure to bilingual or monolingual maternal speech during pregnancy affects the neurophysiological encoding of speech sounds in neonates differently. Front Hum Neurosci 2024; 18:1379660. [PMID: 38841122 PMCID: PMC11150635 DOI: 10.3389/fnhum.2024.1379660] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Accepted: 04/22/2024] [Indexed: 06/07/2024] Open
Abstract
Introduction Exposure to maternal speech during the prenatal period shapes speech perception and linguistic preferences, allowing neonates to recognize stories heard frequently in utero and demonstrating an enhanced preference for their mother's voice and native language. Yet, with a high prevalence of bilingualism worldwide, it remains an open question whether monolingual or bilingual maternal speech during pregnancy influence differently the fetus' neural mechanisms underlying speech sound encoding. Methods In the present study, the frequency-following response (FFR), an auditory evoked potential that reflects the complex spectrotemporal dynamics of speech sounds, was recorded to a two-vowel /oa/ stimulus in a sample of 129 healthy term neonates within 1 to 3 days after birth. Newborns were divided into two groups according to maternal language usage during the last trimester of gestation (monolingual; bilingual). Spectral amplitudes and spectral signal-to-noise ratios (SNR) at the stimulus fundamental (F0) and first formant (F1) frequencies of each vowel were, respectively, taken as measures of pitch and formant structure neural encoding. Results Our results reveal that while spectral amplitudes at F0 did not differ between groups, neonates from bilingual mothers exhibited a lower spectral SNR. Additionally, monolingually exposed neonates exhibited a higher spectral amplitude and SNR at F1 frequencies. Discussion We interpret our results under the consideration that bilingual maternal speech, as compared to monolingual, is characterized by a greater complexity in the speech sound signal, rendering newborns from bilingual mothers more sensitive to a wider range of speech frequencies without generating a particularly strong response at any of them. Our results contribute to an expanding body of research indicating the influence of prenatal experiences on language acquisition and underscore the necessity of including prenatal language exposure in developmental studies on language acquisition, a variable often overlooked yet capable of influencing research outcomes.
Collapse
Affiliation(s)
- Natàlia Gorina-Careta
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
| | - Sonia Arenillas-Alcón
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
| | - Marta Puertollano
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
| | - Alejandro Mondéjar-Segovia
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
| | - Siham Ijjou-Kadiri
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
| | - Jordi Costa-Faidella
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
| | - María Dolores Gómez-Roig
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
- BCNatal – Barcelona Center for Maternal Fetal and Neonatal Medicine (Hospital Sant Joan de Déu and Hospital Clínic), University of Barcelona, Barcelona, Spain
| | - Carles Escera
- Brainlab – Cognitive Neuroscience Research Group, Departament de Psicologia Clinica i Psicobiologia, Universitat de Barcelona, Barcelona, Spain
- Institut de Neurociènces, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, Esplugues de Llobregat, Barcelona, Spain
| |
Collapse
|
2
|
Ulanov M, Kopytin G, Bermúdez-Margaretto B, Ntoumanis I, Gorin A, Moiseenko O, Blagovechtchenski E, Moiseeva V, Shestakova A, Jääskeläinen I, Shtyrov Y. Regionally specific cortical lateralization of abstract and concrete verb processing: Magnetic mismatch negativity study. Neuropsychologia 2024; 195:108800. [PMID: 38246413 DOI: 10.1016/j.neuropsychologia.2024.108800] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Revised: 11/03/2023] [Accepted: 01/14/2024] [Indexed: 01/23/2024]
Abstract
The neural underpinnings of processing concrete and abstract semantics remain poorly understood. Previous fMRI studies have shown that multimodal and amodal neural networks respond differentially to different semantic types; importantly, abstract semantics activates more left-lateralized networks, as opposed to more bilateral activity for concrete words. Due to the lack of temporal resolution, these fMRI results do not allow to easily separate language- and task-specific brain responses and to disentangle early processing stages from later post-comprehension phenomena. To tackle this, we used magnetoencephalography (MEG), a time-resolved neuroimaging technique, in combination with a task-free oddball mismatch negativity (MMN) paradigm, an established approach to tracking early automatic activation of word-specific memory traces in the brain. We recorded the magnetic MMN responses in 30 healthy adults to auditorily presented abstract and concrete action verbs to assess lateralization of word-specific lexico-semantic processing in a set of neocortical areas. We found that MMN responses to these stimuli showed different lateralization patterns of activity in the upper limb motor area (BA4) and parts of Broca's area (BA45/BA47) within ∼100-350 ms after the word disambiguation point. Importantly, the greater leftward response lateralization for abstract semantics was due to the lesser involvement of the right-hemispheric homologues, not increased left-hemispheric activity. These findings suggest differential region-specific involvement of bilateral sensorimotor systems already in the early automatic stages of processing abstract and concrete action semantics.
Collapse
Affiliation(s)
- Maxim Ulanov
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia.
| | - Grigory Kopytin
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Beatriz Bermúdez-Margaretto
- Universidad de Salamanca, Facultad de Psicología, Departamento de Psicología Básica, Psicobiología y Metodología de Las Ciencias Del Comportamiento, Salamanca, Spain; Instituto de Integración en La Comunidad - INICO, Salamanca, Spain
| | - Ioannis Ntoumanis
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Aleksei Gorin
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Olesya Moiseenko
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | | | - Victoria Moiseeva
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Anna Shestakova
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Iiro Jääskeläinen
- HSE University, Institute for Cognitive Neuroscience, Moscow, Russia
| | - Yury Shtyrov
- Center of Functionally Integrative Neuroscience (CFIN), Aarhus University, Aarhus, Denmark
| |
Collapse
|
3
|
Kalchev E. Beyond the Sound Waves: A Comprehensive Exploration of the Burn-In Phenomenon in Audio Equipment Across Physiological, Psychological, and Societal Domains. Cureus 2024; 16:e53097. [PMID: 38414701 PMCID: PMC10898501 DOI: 10.7759/cureus.53097] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/28/2024] [Indexed: 02/29/2024] Open
Abstract
Audio burn-in, often referred to as the process by which audio equipment undergoes a series of played sounds to achieve optimal performance, remains a topic of significant debate within both audiophile communities and relevant scientific fields. While some attribute perceived changes in sound quality to actual physical changes in the equipment, an emerging perspective points to the interplay of physiological, psychological, and social factors that might influence these perceptions. This narrative review delves into the intricate layers of auditory physiology, cognitive sound interpretation, and the wider societal beliefs around burn-in. We underscore the importance of discerning between actual physical changes in audio gear and the multifaceted human factors that potentially modulate our perception of sound. Through a comprehensive exploration, this article illuminates the complexities of this phenomenon, offering insights for both medical professionals and passionate audio enthusiasts and proposing directions for future research.
Collapse
Affiliation(s)
- Emilian Kalchev
- Diagnostic Imaging, St. Marina University Hospital, Varna, BGR
| |
Collapse
|
4
|
Bellur A, Thakkar K, Elhilali M. Explicit-memory multiresolution adaptive framework for speech and music separation. EURASIP JOURNAL ON AUDIO, SPEECH, AND MUSIC PROCESSING 2023; 2023:20. [PMID: 37181589 PMCID: PMC10169896 DOI: 10.1186/s13636-023-00286-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Accepted: 04/21/2023] [Indexed: 05/16/2023]
Abstract
The human auditory system employs a number of principles to facilitate the selection of perceptually separated streams from a complex sound mixture. The brain leverages multi-scale redundant representations of the input and uses memory (or priors) to guide the selection of a target sound from the input mixture. Moreover, feedback mechanisms refine the memory constructs resulting in further improvement of selectivity of a particular sound object amidst dynamic backgrounds. The present study proposes a unified end-to-end computational framework that mimics these principles for sound source separation applied to both speech and music mixtures. While the problems of speech enhancement and music separation have often been tackled separately due to constraints and specificities of each signal domain, the current work posits that common principles for sound source separation are domain-agnostic. In the proposed scheme, parallel and hierarchical convolutional paths map input mixtures onto redundant but distributed higher-dimensional subspaces and utilize the concept of temporal coherence to gate the selection of embeddings belonging to a target stream abstracted in memory. These explicit memories are further refined through self-feedback from incoming observations in order to improve the system's selectivity when faced with unknown backgrounds. The model yields stable outcomes of source separation for both speech and music mixtures and demonstrates benefits of explicit memory as a powerful representation of priors that guide information selection from complex inputs.
Collapse
Affiliation(s)
- Ashwin Bellur
- Electrical and Computer Engineering, Johns Hopkins University, Baltimore, USA
| | - Karan Thakkar
- Electrical and Computer Engineering, Johns Hopkins University, Baltimore, USA
| | - Mounya Elhilali
- Electrical and Computer Engineering, Johns Hopkins University, Baltimore, USA
| |
Collapse
|
5
|
Leong ATL, Wong EC, Wang X, Wu EX. Hippocampus Modulates Vocalizations Responses at Early Auditory Centers. Neuroimage 2023; 270:119943. [PMID: 36828157 DOI: 10.1016/j.neuroimage.2023.119943] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2023] [Accepted: 02/13/2023] [Indexed: 02/24/2023] Open
Abstract
Despite its prominence in learning and memory, hippocampal influence in early auditory processing centers remains unknown. Here, we examined how hippocampal activity modulates sound-evoked responses in the auditory midbrain and thalamus using optogenetics and functional MRI (fMRI) in rodents. Ventral hippocampus (vHP) excitatory neuron stimulation at 5 Hz evoked robust hippocampal activity that propagates to the primary auditory cortex. We then tested 5 Hz vHP stimulation paired with either natural vocalizations or artificial/noise acoustic stimuli. vHP stimulation enhanced auditory responses to vocalizations (with a negative or positive valence) in the inferior colliculus, medial geniculate body, and auditory cortex, but not to their temporally reversed counterparts (artificial sounds) or broadband noise. Meanwhile, pharmacological vHP inactivation diminished response selectivity to vocalizations. These results directly reveal the large-scale hippocampal participation in natural sound processing at early centers of the ascending auditory pathway. They expand our present understanding of hippocampus in global auditory networks.
Collapse
Affiliation(s)
- Alex T L Leong
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Pokfulam, Hong Kong SAR, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam, Hong Kong SAR, China.
| | - Eddie C Wong
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Pokfulam, Hong Kong SAR, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam, Hong Kong SAR, China
| | - Xunda Wang
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Pokfulam, Hong Kong SAR, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam, Hong Kong SAR, China
| | - Ed X Wu
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Pokfulam, Hong Kong SAR, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam, Hong Kong SAR, China; School of Biomedical Sciences, LKS Faculty of Medicine, The University of Hong Kong, Pokfulam, Hong Kong SAR, China.
| |
Collapse
|
6
|
Zhou B, Tomioka R, Song WJ. Temporal profiles of neuronal responses to repeated tone stimuli in the mouse primary auditory cortex. Hear Res 2023; 430:108710. [PMID: 36758331 DOI: 10.1016/j.heares.2023.108710] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/14/2022] [Revised: 01/26/2023] [Accepted: 02/01/2023] [Indexed: 02/05/2023]
Abstract
How the auditory system processes temporal information of sound has been investigated extensively using repeated stimuli. Recent studies on how the response of neurons in the primary auditory cortex (A1) changes with the progression of stimulus repetition, have reported response temporal profiles of two categories: "adaptation", i.e., gradual decrease, and "facilitation", i.e., gradual increase. To explore the existence of profiles of other categories and to examine the tone-frequency-dependence of the profile category in single neurons, here we studied the response of mouse A1 neurons to four or five tone-trains; each train comprised 10 identical tone pips, with 0.5-s inter-tone-intervals, and the four or five trains differed only in tone frequency. The response to each tone in a train was evaluated using the peak of the ON response, and how the peak response changed with the tone number in the train, i.e., the response temporal profile, was examined. We confirmed the existence of profiles of both "adaptation" and "facilitation" categories; "adaptation" could be further subcategorized into "slow adaptation" and "fast adaptation" profiles, with the latter being encountered more frequently. Moreover, two new categories of non-monotonic profiles were identified: an "adaptation with recovery" profile and a "facilitation followed by adaptation" profile. Examination of single neurons with trains of different tone frequencies revealed that some A1 neurons exhibited profiles of the same category to tone trains of different tone frequencies, whereas others exhibited profiles of different categories, depending on the tone frequency. These results demonstrate the variety in the response temporal profiles of mouse A1 neurons, which may benefit the encoding of individual tones in a train.
Collapse
Affiliation(s)
- Bo Zhou
- Department of Sensory and Cognitive Physiology, Graduate School of Medical Sciences, Kumamoto University 860-8556, Japan
| | - Ryohei Tomioka
- Department of Sensory and Cognitive Physiology, Graduate School of Medical Sciences, Kumamoto University 860-8556, Japan.
| | - Wen-Jie Song
- Department of Sensory and Cognitive Physiology, Graduate School of Medical Sciences, Kumamoto University 860-8556, Japan; Center for Metabolic Regulation of Healthy Aging, Faculty of Life Sciences, Kumamoto University, Kumamoto 860-8556, Japan.
| |
Collapse
|
7
|
Weise A, Grimm S, Maria Rimmele J, Schröger E. Auditory representations for long lasting sounds: Insights from event-related brain potentials and neural oscillations. BRAIN AND LANGUAGE 2023; 237:105221. [PMID: 36623340 DOI: 10.1016/j.bandl.2022.105221] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/23/2021] [Revised: 12/26/2022] [Accepted: 12/27/2022] [Indexed: 06/17/2023]
Abstract
The basic features of short sounds, such as frequency and intensity including their temporal dynamics, are integrated in a unitary representation. Knowledge on how our brain processes long lasting sounds is scarce. We review research utilizing the Mismatch Negativity event-related potential and neural oscillatory activity for studying representations for long lasting simple versus complex sounds such as sinusoidal tones versus speech. There is evidence for a temporal constraint in the formation of auditory representations: Auditory edges like sound onsets within long lasting sounds open a temporal window of about 350 ms in which the sounds' dynamics are integrated into a representation, while information beyond that window contributes less to that representation. This integration window segments the auditory input into short chunks. We argue that the representations established in adjacent integration windows can be concatenated into an auditory representation of a long sound, thus, overcoming the temporal constraint.
Collapse
Affiliation(s)
- Annekathrin Weise
- Department of Psychology, Ludwig-Maximilians-University Munich, Germany; Wilhelm Wundt Institute for Psychology, Leipzig University, Germany.
| | - Sabine Grimm
- Wilhelm Wundt Institute for Psychology, Leipzig University, Germany.
| | - Johanna Maria Rimmele
- Department of Neuroscience, Max-Planck-Institute for Empirical Aesthetics, Germany; Center for Language, Music and Emotion, New York University, Max Planck Institute, Department of Psychology, 6 Washington Place, New York, NY 10003, United States.
| | - Erich Schröger
- Wilhelm Wundt Institute for Psychology, Leipzig University, Germany.
| |
Collapse
|
8
|
Tuned in to communication sounds: Neuronal sensitivity in the túngara frog midbrain to frequency modulated signals. PLoS One 2022; 17:e0268383. [PMID: 35587486 PMCID: PMC9119527 DOI: 10.1371/journal.pone.0268383] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2021] [Accepted: 04/28/2022] [Indexed: 11/19/2022] Open
Abstract
For complex communication signals, it is often difficult to identify the information-bearing elements and their parameters necessary to elicit functional behavior. Consequently, it may be difficult to design stimuli that test how neurons contribute to communicative processing. For túngara frogs (Physalaemus pustulosus), however, previous behavioral testing with numerous stimuli showed that a particular frequency modulated (FM) transition in the male call is required to elicit phonotaxis and vocal responses. Modeled on such behavioral experiments, we used awake in vivo recordings of single units in the midbrain to determine if their excitation was biased to behaviorally important FM parameters. Comparisons of stimulus driven action potentials revealed greatest excitation to the behaviorally important FM transition: a downward FM sweep or step that crosses ~600 Hz. Previous studies using long-duration acoustic exposure found immediate early gene expression in many midbrain neurons to be most sensitive to similar FM. However, those data could not determine if FM coding was accomplished by the population and/or individual neurons. Our data suggest both coding schemes could operate, as 1) individual neurons are more sensitive to the behaviorally significant FM transition and 2) when single unit recordings are analytically combined across cells, the combined code can produce high stimulus discrimination (FM vs. noise driven excitation), approaching that found in behavioral discrimination of call vs. noise.
Collapse
|
9
|
Liu D, Hu J, Wang S, Fu X, Wang Y, Pugh E, Henderson Sabes J, Wang S. Aging Affects Subcortical Pitch Information Encoding Differently in Humans With Different Language Backgrounds. Front Aging Neurosci 2022; 14:816100. [PMID: 35493942 PMCID: PMC9043765 DOI: 10.3389/fnagi.2022.816100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Accepted: 03/16/2022] [Indexed: 11/13/2022] Open
Abstract
Aging and language background have been shown to affect pitch information encoding at the subcortical level. To study the individual and compounded effects on subcortical pitch information encoding, Frequency Following Responses were recorded from subjects across various ages and language backgrounds. Differences were found in pitch information encoding strength and accuracy among the groups, indicating that language experience and aging affect accuracy and magnitude of pitch information encoding ability at the subcortical level. Moreover, stronger effects of aging were seen in the magnitude of phase-locking in the native language speaker groups, while language background appears to have more impact on the accuracy of pitch tracking in older adult groups.
Collapse
Affiliation(s)
- Dongxin Liu
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Institute of Otolaryngology, Otolaryngology—Head and Neck Surgery, Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| | - Jiong Hu
- Department of Audiology, University of the Pacific, San Francisco, CA, United States
| | - Songjian Wang
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Institute of Otolaryngology, Otolaryngology—Head and Neck Surgery, Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| | - Xinxing Fu
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Institute of Otolaryngology, Otolaryngology—Head and Neck Surgery, Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| | - Yuan Wang
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Institute of Otolaryngology, Otolaryngology—Head and Neck Surgery, Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| | - Esther Pugh
- Department of Otolaryngology, Keck School of Medicine of USC, Los Angeles, CA, United States
| | | | - Shuo Wang
- Key Laboratory of Otolaryngology Head and Neck Surgery, Beijing Institute of Otolaryngology, Otolaryngology—Head and Neck Surgery, Ministry of Education, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| |
Collapse
|
10
|
Relationship between objective measures of hearing discrimination elicited by non-linguistic stimuli and speech perception in adults. Sci Rep 2021; 11:19554. [PMID: 34599244 PMCID: PMC8486784 DOI: 10.1038/s41598-021-98950-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2021] [Accepted: 09/14/2021] [Indexed: 11/08/2022] Open
Abstract
Some people using hearing aids have difficulty discriminating between sounds even though the sounds are audible. As such, cochlear implants may provide greater benefits for speech perception. One method to identify people with auditory discrimination deficits is to measure discrimination thresholds using spectral ripple noise (SRN). Previous studies have shown that behavioral discrimination of SRN was associated with speech perception, and behavioral discrimination was also related to cortical responses to acoustic change or ACCs. We hypothesized that cortical ACCs could be directly related to speech perception. In this study, we investigated the relationship between subjective speech perception and objective ACC responses measured using SRNs. We tested 13 normal-hearing and 10 hearing-impaired adults using hearing aids. Our results showed that behavioral SRN discrimination was correlated with speech perception in quiet and in noise. Furthermore, cortical ACC responses to phase changes in the SRN were significantly correlated with speech perception. Audibility was a major predictor of discrimination and speech perception, but direct measures of auditory discrimination could contribute information about a listener’s sensitivity to acoustic cues that underpin speech perception. The findings lend support for potential application of measuring ACC responses to SRNs for identifying people who may benefit from cochlear implants.
Collapse
|
11
|
Rutherford MA, von Gersdorff H, Goutman JD. Encoding sound in the cochlea: from receptor potential to afferent discharge. J Physiol 2021; 599:2527-2557. [PMID: 33644871 PMCID: PMC8127127 DOI: 10.1113/jp279189] [Citation(s) in RCA: 23] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2020] [Accepted: 02/22/2021] [Indexed: 12/17/2022] Open
Abstract
Ribbon-class synapses in the ear achieve analog to digital transformation of a continuously graded membrane potential to all-or-none spikes. In mammals, several auditory nerve fibres (ANFs) carry information from each inner hair cell (IHC) to the brain in parallel. Heterogeneity of transmission among synapses contributes to the diversity of ANF sound-response properties. In addition to the place code for sound frequency and the rate code for sound level, there is also a temporal code. In series with cochlear amplification and frequency tuning, neural representation of temporal cues over a broad range of sound levels enables auditory comprehension in noisy multi-speaker settings. The IHC membrane time constant introduces a low-pass filter that attenuates fluctuations of the receptor potential above 1-2 kHz. The ANF spike generator adds a high-pass filter via its depolarization-rate threshold that rejects slow changes in the postsynaptic potential and its phasic response property that ensures one spike per depolarization. Synaptic transmission involves several stochastic subcellular processes between IHC depolarization and ANF spike generation, introducing delay and jitter that limits the speed and precision of spike timing. ANFs spike at a preferred phase of periodic sounds in a process called phase-locking that is limited to frequencies below a few kilohertz by both the IHC receptor potential and the jitter in synaptic transmission. During phase-locking to periodic sounds of increasing intensity, faster and facilitated activation of synaptic transmission and spike generation may be offset by presynaptic depletion of synaptic vesicles, resulting in relatively small changes in response phase. Here we review encoding of spike-timing at cochlear ribbon synapses.
Collapse
Affiliation(s)
- Mark A. Rutherford
- Department of Otolaryngology, Washington University School of Medicine, St. Louis, Missouri 63110
| | - Henrique von Gersdorff
- Vollum Institute, Oregon Hearing Research Center, Oregon Health and Sciences University, Portland, Oregon 97239
| | | |
Collapse
|
12
|
Neural encoding of voice pitch and formant structure at birth as revealed by frequency-following responses. Sci Rep 2021; 11:6660. [PMID: 33758251 PMCID: PMC7987955 DOI: 10.1038/s41598-021-85799-x] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2020] [Accepted: 03/04/2021] [Indexed: 11/22/2022] Open
Abstract
Detailed neural encoding of voice pitch and formant structure plays a crucial role in speech perception, and is of key importance for an appropriate acquisition of the phonetic repertoire in infants since birth. However, the extent to what newborns are capable of extracting pitch and formant structure information from the temporal envelope and the temporal fine structure of speech sounds, respectively, remains unclear. Here, we recorded the frequency-following response (FFR) elicited by a novel two-vowel, rising-pitch-ending stimulus to simultaneously characterize voice pitch and formant structure encoding accuracy in a sample of neonates and adults. Data revealed that newborns tracked changes in voice pitch reliably and no differently than adults, but exhibited weaker signatures of formant structure encoding, particularly at higher formant frequency ranges. Thus, our results indicate a well-developed encoding of voice pitch at birth, while formant structure representation is maturing in a frequency-dependent manner. Furthermore, we demonstrate the feasibility to assess voice pitch and formant structure encoding within clinical evaluation times in a hospital setting, and suggest the possibility to use this novel stimulus as a tool for longitudinal developmental studies of the auditory system.
Collapse
|
13
|
Fox NP, Leonard M, Sjerps MJ, Chang EF. Transformation of a temporal speech cue to a spatial neural code in human auditory cortex. eLife 2020; 9:e53051. [PMID: 32840483 PMCID: PMC7556862 DOI: 10.7554/elife.53051] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2019] [Accepted: 08/21/2020] [Indexed: 11/28/2022] Open
Abstract
In speech, listeners extract continuously-varying spectrotemporal cues from the acoustic signal to perceive discrete phonetic categories. Spectral cues are spatially encoded in the amplitude of responses in phonetically-tuned neural populations in auditory cortex. It remains unknown whether similar neurophysiological mechanisms encode temporal cues like voice-onset time (VOT), which distinguishes sounds like /b/ and/p/. We used direct brain recordings in humans to investigate the neural encoding of temporal speech cues with a VOT continuum from /ba/ to /pa/. We found that distinct neural populations respond preferentially to VOTs from one phonetic category, and are also sensitive to sub-phonetic VOT differences within a population's preferred category. In a simple neural network model, simulated populations tuned to detect either temporal gaps or coincidences between spectral cues captured encoding patterns observed in real neural data. These results demonstrate that a spatial/amplitude neural code underlies the cortical representation of both spectral and temporal speech cues.
Collapse
Affiliation(s)
- Neal P Fox
- Department of Neurological Surgery, University of California, San FranciscoSan FranciscoUnited States
| | - Matthew Leonard
- Department of Neurological Surgery, University of California, San FranciscoSan FranciscoUnited States
| | - Matthias J Sjerps
- Donders Institute for Brain, Cognition and Behaviour, Centre for Cognitive Neuroimaging, Radboud UniversityNijmegenNetherlands
- Max Planck Institute for PsycholinguisticsNijmegenNetherlands
| | - Edward F Chang
- Department of Neurological Surgery, University of California, San FranciscoSan FranciscoUnited States
- Weill Institute for Neurosciences, University of California, San FranciscoSan FranciscoUnited States
| |
Collapse
|
14
|
Eggermont JJ. Separate auditory pathways for the induction and maintenance of tinnitus and hyperacusis? PROGRESS IN BRAIN RESEARCH 2020; 260:101-127. [PMID: 33637214 DOI: 10.1016/bs.pbr.2020.01.006] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
Abstract
Tinnitus and hyperacusis often occur together, however tinnitus may occur without hyperacusis or hyperacusis without tinnitus. Based on animal research one could argue that hyperacusis results from noise exposures that increase central gain in the lemniscal, tonotopically organized, pathways, whereas tinnitus requires increased burst firing and neural synchrony in the extra-lemniscal pathway. However, these substrates are not sufficient and require involvement of the central nervous system. The dominant factors in changing cortical networks in tinnitus patients are foremost the degree and type of hearing loss, and comorbidities such as distress and mood. So far, no definite changes have been established for tinnitus proper, albeit that changes in connectivity between the dorsal attention network and the parahippocampal area, as well as the default-mode network-precuneus decoupling, appear to be strong candidates. I conclude that there is still a strong need for further integrating animal and human research into tinnitus and hyperacusis.
Collapse
Affiliation(s)
- Jos J Eggermont
- Department of Psychology, Department of Physiology and Pharmacology, University of Calgary, Calgary, AB, Canada.
| |
Collapse
|
15
|
Bellur A, Elhilali M. Audio object classification using distributed beliefs and attention. IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING 2020; 28:729-739. [PMID: 33564695 PMCID: PMC7869589 DOI: 10.1109/taslp.2020.2966867] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
One of the unique characteristics of human hearing is its ability to recognize acoustic objects even in presence of severe noise and distortions. In this work, we explore two mechanisms underlying this ability: 1) redundant mapping of acoustic waveforms along distributed latent representations and 2) adaptive feedback based on prior knowledge to selectively attend to targets of interest. We propose a bio-mimetic account of acoustic object classification by developing a novel distributed deep belief network validated for the task of robust acoustic object classification using the UrbanSound database. The proposed distributed belief network (DBN) encompasses an array of independent sub-networks trained generatively to capture different abstractions of natural sounds. A supervised classifier then performs a readout of this distributed mapping. The overall architecture not only matches the state of the art system for acoustic object classification but leads to significant improvement over the baseline in mismatched noisy conditions (31.4% relative improvement in 0dB conditions). Furthermore, we incorporate mechanisms of attentional feedback that allows the DBN to deploy local memories of sounds targets estimated at multiple views to bias network activation when attending to a particular object. This adaptive feedback results in further improvement of object classification in unseen noise conditions (relative improvement of 54% over the baseline in 0dB conditions).
Collapse
Affiliation(s)
- Ashwin Bellur
- Department of Electrical and Computer Engineering, Laboratory for Computational Audio Perception, Johns Hopkins University
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Laboratory for Computational Audio Perception, Johns Hopkins University
| |
Collapse
|
16
|
Abstract
OBJECTIVES The objectives of this study were to measure the effects of level and vowel contrast on the latencies and amplitudes of acoustic change complex (ACC) in the mature auditory system. This was done to establish how the ACC in healthy young adults is affected by these stimulus parameters that could then be used to inform translation of the ACC into a clinical measure for the pediatric population. Another aim was to demonstrate that a normalized amplitude metric, calculated by dividing the ACC amplitude in the vowel contrast condition by the ACC amplitude obtained in a control condition (no vowel change) would demonstrate good sensitivity with respect to perceptual measures of vowel-contrast detection. The premises underlying this research were that: (1) ACC latencies and amplitudes would vary with level, in keeping with principles of an increase in neural synchrony and activity that takes place as a function of increasing stimulus level; (2) ACC latencies and amplitudes would vary with vowel contrast, because cortical auditory evoked potentials are known to be sensitive to the spectro-temporal characteristics of speech. DESIGN Nineteen adults, 14 of them female, with a mean age of 24.2 years (range 20 to 38 years) participated in this study. All had normal-hearing thresholds. Cortical auditory evoked potentials were obtained from all participants in response to synthesized vowel tokens (/a/, /i/, /o/, /u/), presented in a quasi-steady state fashion at a rate of 2/sec in an oddball stimulus paradigm, with a 25% probability of the deviant stimulus. The ACC was obtained in response to the deviant stimulus. All combinations of vowel tokens were tested at 2 stimulus levels: 40 and 70 dBA. In addition, listeners were tested for their ability to detect the vowel contrasts using behavioral methods. RESULTS ACC amplitude varied systematically with level, and test condition (control versus contrast) and vowel token, but ACC latency did not. ACC amplitudes were significantly larger when tested at 70 dBA compared with 40 dBA and for contrast trials compared with control trials at both levels. Amplitude ratios (normalized amplitudes) were largest for contrast pairs in which /a/ was the standard token. The amplitude ratio metric at the individual level demonstrated up to 97% sensitivity with respect to perceptual measures of discrimination. CONCLUSIONS The present study establishes the effects of stimulus level and vowel type on the latency and amplitude of the ACC in the young adult auditory system and supports the amplitude ratio as a sensitive metric for cortical acoustic salience of vowel spectral features. Next steps are to evaluate these methods in infants and children with hearing loss with the long-term goal of its translation into a clinical method for estimating speech feature discrimination.
Collapse
|
17
|
Ross B, Tremblay KL, Alain C. Simultaneous EEG and MEG recordings reveal vocal pitch elicited cortical gamma oscillations in young and older adults. Neuroimage 2019; 204:116253. [PMID: 31600592 DOI: 10.1016/j.neuroimage.2019.116253] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2019] [Revised: 09/13/2019] [Accepted: 10/06/2019] [Indexed: 10/25/2022] Open
Abstract
The frequency-following response with origin in the auditory brainstem represents the pitch contour of voice and can be recorded with electrodes from the scalp. MEG studies also revealed a cortical contribution to the high gamma oscillations at the fundamental frequency (f0) of a vowel stimulus. Therefore, studying the cortical component of the frequency-following response could provide insights into how pitch information is encoded at the cortical level. Comparing how aging affects the different responses may help to uncover the neural mechanisms underlying speech understanding deficits in older age. We simultaneously recorded EEG and MEG responses to the syllable /ba/. MEG beamformer analysis localized sources in bilateral auditory cortices and the midbrain. Time-frequency analysis showed a faithful representation of the pitch contour between 106 Hz and 138 Hz in the cortical activity. A cross-correlation revealed a latency of 20 ms. Furthermore, stimulus onsets elicited cortical 40-Hz responses. Both the 40-Hz and the f0 response amplitudes increased in older age and were larger in the right hemisphere. The effects of aging and laterality of the f0 response were evident in the MEG only, suggesting that both effects were characteristics of the cortical response. After comparing f0 and N1 responses in EEG and MEG, we estimated that approximately one-third of the scalp-recorded f0 response could be cortical in origin. We attributed the significance of the cortical f0 response to the precise timing of cortical neurons that serve as a time-sensitive code for pitch.
Collapse
Affiliation(s)
- Bernhard Ross
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department for Medical Biophysics, University of Toronto, Ontario, Canada.
| | - Kelly L Tremblay
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada
| |
Collapse
|
18
|
Burghard A, Voigt MB, Kral A, Hubka P. Categorical processing of fast temporal sequences in the guinea pig auditory brainstem. Commun Biol 2019; 2:265. [PMID: 31341964 PMCID: PMC6642126 DOI: 10.1038/s42003-019-0472-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2018] [Accepted: 05/23/2019] [Indexed: 11/21/2022] Open
Abstract
Discrimination of temporal sequences is crucial for auditory object recognition, phoneme categorization and speech understanding. The present study shows that auditory brainstem responses (ABR) to pairs of noise bursts separated by a short gap can be classified into two distinct groups based on the ratio of gap duration to initial noise burst duration in guinea pigs. If this ratio was smaller than 0.5, the ABR to the trailing noise burst was strongly suppressed. On the other hand, if the initial noise burst duration was short compared to the gap duration (a ratio greater than 0.5), a release from suppression and/or enhancement of the trailing ABR was observed. Consequently, initial noise bursts of shorter duration caused a faster transition between response classes than initial noise bursts of longer duration. We propose that the described findings represent a neural correlate of subcortical categorical preprocessing of temporal sequences in the auditory system.
Collapse
Affiliation(s)
- Alice Burghard
- Institute of Audioneurotechnology & Department of Experimental Otology, ENT Clinics, Hannover Medical School, Hannover, D-30625 Germany
- Department of Neuroscience, University of Connecticut Health Center, Farmington, CT 06030 USA
| | - Mathias Benjamin Voigt
- Institute of Audioneurotechnology & Department of Experimental Otology, ENT Clinics, Hannover Medical School, Hannover, D-30625 Germany
| | - Andrej Kral
- Institute of Audioneurotechnology & Department of Experimental Otology, ENT Clinics, Hannover Medical School, Hannover, D-30625 Germany
| | - Peter Hubka
- Institute of Audioneurotechnology & Department of Experimental Otology, ENT Clinics, Hannover Medical School, Hannover, D-30625 Germany
| |
Collapse
|
19
|
Felix RA, Chavez VA, Novicio DM, Morley BJ, Portfors CV. Nicotinic acetylcholine receptor subunit α 7-knockout mice exhibit degraded auditory temporal processing. J Neurophysiol 2019; 122:451-465. [PMID: 31116647 DOI: 10.1152/jn.00170.2019] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/05/2023] Open
Abstract
The CHRNA7 gene that encodes the α7-subunit of the nicotinic acetylcholine receptor (α7-nAChR) has been associated with some autism spectrum disorders and other neurodevelopmental conditions characterized, in part, by auditory and language impairment. These conditions may include auditory processing disorders that represent impaired timing of neural activity, often accompanied by problems understanding speech. Here, we measure timing properties of sound-evoked activity via the auditory brainstem response (ABR) of α7-nAChR knockout mice of both sexes and wild-type colony controls. We find a significant timing delay in evoked ABR signals that represents midbrain activity in knockouts. We also examine spike-timing properties of neurons in the inferior colliculus, a midbrain nucleus that exhibits high levels of α7-nAChR during development. We find delays of evoked responses along with degraded spiking precision in knockout animals. We find similar timing deficits in responses of neurons in the superior paraolivary nucleus and ventral nucleus of the lateral lemniscus, which are brainstem nuclei thought to shape temporal precision in the midbrain. In addition, we find that other measures of temporal acuity including forward masking and gap detection are impaired for knockout animals. We conclude that altered temporal processing at the level of the brainstem in α7-nAChR-deficient mice may contribute to degraded spike timing in the midbrain, which may underlie the observed timing delay in the ABR signals. Our findings are consistent with a role for the α7-nAChR in types of neurodevelopmental and auditory processing disorders and we identify potential neural targets for intervention.NEW & NOTEWORTHY Disrupted signaling via the α7-nicotinic acetylcholine receptor (α7-nAChR) is associated with neurodevelopmental disorders that include impaired auditory processing. The underlying causes of dysfunction are not known but a common feature is abnormal timing of neural activity. We examined temporal processing of α7-nAChR knockout mice and wild-type controls. We found degraded spike timing of neurons in knockout animals, which manifests at the level of the auditory brainstem and midbrain.
Collapse
Affiliation(s)
- Richard A Felix
- School of Biological Sciences and the Department of Integrated Physiology and Neuroscience, Washington State University Vancouver, Vancouver, Washington
| | - Vicente A Chavez
- School of Biological Sciences and the Department of Integrated Physiology and Neuroscience, Washington State University Vancouver, Vancouver, Washington
| | - Dyana M Novicio
- School of Biological Sciences and the Department of Integrated Physiology and Neuroscience, Washington State University Vancouver, Vancouver, Washington
| | | | - Christine V Portfors
- School of Biological Sciences and the Department of Integrated Physiology and Neuroscience, Washington State University Vancouver, Vancouver, Washington
| |
Collapse
|
20
|
Roque L, Gaskins C, Gordon-Salant S, Goupell MJ, Anderson S. Age Effects on Neural Representation and Perception of Silence Duration Cues in Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:1099-1116. [PMID: 31026197 PMCID: PMC6802877 DOI: 10.1044/2018_jslhr-h-ascc7-18-0076] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/05/2018] [Revised: 06/26/2018] [Accepted: 08/12/2018] [Indexed: 06/09/2023]
Abstract
Purpose Degraded temporal processing associated with aging may be a contributing factor to older adults' hearing difficulties, especially in adverse listening environments. This degraded processing may affect the ability to distinguish between words based on temporal duration cues. The current study investigates the effects of aging and hearing loss on cortical and subcortical representation of temporal speech components and on the perception of silent interval duration cues in speech. Method Identification functions for the words DISH and DITCH were obtained on a 7-step continuum of silence duration (0-60 ms) prior to the final fricative in participants who are younger with normal hearing (YNH), older with normal hearing (ONH), and older with hearing impairment (OHI). Frequency-following responses and cortical auditory-evoked potentials were recorded to the 2 end points of the continuum. Auditory brainstem responses to clicks were obtained to verify neural integrity and to compare group differences in auditory nerve function. A multiple linear regression analysis was conducted to determine the peripheral or central factors that contributed to perceptual performance. Results ONH and OHI participants required longer silence durations to identify DITCH than did YNH participants. Frequency-following responses showed reduced phase locking and poorer morphology, and cortical auditory-evoked potentials showed prolonged latencies in ONH and OHI participants compared with YNH participants. No group differences were noted for auditory brainstem response Wave I amplitude or Wave V/I ratio. After accounting for the possible effects of hearing loss, linear regression analysis revealed that both midbrain and cortical processing contributed to the variance in the DISH-DITCH perceptual identification functions. Conclusions These results suggest that age-related deficits in the ability to encode silence duration cues may be a contributing factor in degraded speech perception. In particular, degraded response morphology relates to performance on perceptual tasks based on silence duration contrasts between words.
Collapse
Affiliation(s)
- Lindsey Roque
- Department of Hearing and Speech Sciences, University of Maryland, College Park
| | - Casey Gaskins
- Department of Hearing and Speech Sciences, University of Maryland, College Park
| | - Sandra Gordon-Salant
- Department of Hearing and Speech Sciences, University of Maryland, College Park
- Neuroscience and Cognitive Science Program, University of Maryland, College Park
| | - Matthew J. Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park
- Neuroscience and Cognitive Science Program, University of Maryland, College Park
| | - Samira Anderson
- Department of Hearing and Speech Sciences, University of Maryland, College Park
- Neuroscience and Cognitive Science Program, University of Maryland, College Park
| |
Collapse
|
21
|
Zhang F, Underwood G, McGuire K, Liang C, Moore DR, Fu QJ. Frequency change detection and speech perception in cochlear implant users. Hear Res 2019; 379:12-20. [PMID: 31035223 DOI: 10.1016/j.heares.2019.04.007] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/03/2018] [Revised: 03/21/2019] [Accepted: 04/15/2019] [Indexed: 10/27/2022]
Abstract
Dynamic frequency changes in sound provide critical cues for speech perception. Most previous studies examining frequency discrimination in cochlear implant (CI) users have employed behavioral tasks in which target and reference tones (differing in frequency) are presented statically in separate time intervals. Participants are required to identify the target frequency by comparing stimuli across these time intervals. However, perceiving dynamic frequency changes in speech requires detection of within-interval frequency change. This study explored the relationship between detection of within-interval frequency changes and speech perception performance of CI users. Frequency change detection thresholds (FCDTs) were measured in 20 adult CI users using a 3-alternative forced-choice (3AFC) procedure. Stimuli were 1-sec pure tones (base frequencies at 0.25, 1, 4 kHz) with frequency changes occurring 0.5 s after the tone onset. Speech tests were 1) Consonant-Nucleus-Consonant (CNC) monosyllabic word recognition, 2) Arizona Biomedical Sentence Recognition (AzBio) in Quiet, 3) AzBio in Noise (AzBio-N, +10 dB signal-to-noise/SNR ratio), and 4) Digits-in-noise (DIN). Participants' subjective satisfaction with the CI was obtained. Results showed that correlations between FCDTs and speech perception were all statistically significant. The satisfaction level of CI use was not related to FCDTs, after controlling for major demographic factors. DIN speech reception thresholds were significantly correlated to AzBio-N scores. The current findings suggest that the ability to detect within-interval frequency changes may play an important role in speech perception performance of CI users. FCDT and DIN can serve as simple and rapid tests that require no or minimal linguistic background for the prediction of CI speech outcomes.
Collapse
Affiliation(s)
- Fawen Zhang
- Department of Communication Sciences and Disorders, University of Cincinnati, Ohio, USA.
| | - Gabrielle Underwood
- Department of Communication Sciences and Disorders, University of Cincinnati, Ohio, USA
| | - Kelli McGuire
- Department of Communication Sciences and Disorders, University of Cincinnati, Ohio, USA
| | - Chun Liang
- Department of Communication Sciences and Disorders, University of Cincinnati, Ohio, USA; Shenzhen Maternity & Child Healthcare Hospital, Shenzhen, China
| | - David R Moore
- Communication Sciences Research Center, Cincinnati Children's Hospital Medical Center, Cincinnati, OH, USA; Department of Otolaryngology, University of Cincinnati, Ohio, USA
| | - Qian-Jie Fu
- Department of Head and Neck Surgery, University of California, Los Angeles, Los Angeles, CA, USA
| |
Collapse
|
22
|
Zhu S, Allitt B, Samuel A, Lui L, Rosa MGP, Rajan R. Sensitivity to Vocalization Pitch in the Caudal Auditory Cortex of the Marmoset: Comparison of Core and Belt Areas. Front Syst Neurosci 2019; 13:5. [PMID: 30774587 PMCID: PMC6367263 DOI: 10.3389/fnsys.2019.00005] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2018] [Accepted: 01/14/2019] [Indexed: 11/13/2022] Open
Abstract
Based on anatomical connectivity and basic response characteristics, primate auditory cortex is divided into a central core surrounded by belt and parabelt regions. The encoding of pitch, a prototypical element of sound identity, has been studied in primary auditory cortex (A1) but little is known about how it is encoded and represented beyond A1. The caudal auditory belt and parabelt cortical fields process spatial information but also contain information on non-spatial aspects of sounds. In this study, we examined neuronal responses in these areas to pitch-varied marmoset vocalizations, to derive the consequent representation of pitch in these regions and the potential underlying mechanisms, to compare to the encoding and representation of pitch of the same sounds in A1. With respect to response patterns to the vocalizations, neurons in caudal medial belt (CM) showed similar short-latency and short-duration response patterns to A1, but caudal lateral belt (CL) neurons at the same hierarchical level and caudal parabelt (CPB) neurons at a higher hierarchical level showed delayed or much delayed response onset and prolonged response durations. With respect to encoding of pitch, neurons in all cortical fields showed sensitivity to variations in the vocalization pitch either through modulation of spike-count or of first spike-latency. The utility of the encoding mechanism differed between fields: pitch sensitivity was reliably represented by spike-count variations in A1 and CM, while first spike-latency variation was better for encoding pitch in CL and CPB. In summary, our data show that (a) the traditionally-defined belt area CM is functionally very similar to A1 with respect to the representation and encoding of complex naturalistic sounds, (b) the CL belt area, at the same hierarchical level as CM, and the CPB area, at a higher hierarchical level, have very different response patterns and appear to use different pitch-encoding mechanisms, and (c) caudal auditory fields, proposed to be specialized for encoding spatial location, can also contain robust representations of sound identity.
Collapse
Affiliation(s)
- Shuyu Zhu
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia.,Australian Research Council, Centre of Excellence in Integrative Brain Function, Clayton, VIC, Australia
| | - Benjamin Allitt
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia
| | - Anil Samuel
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia
| | - Leo Lui
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia.,Australian Research Council, Centre of Excellence in Integrative Brain Function, Clayton, VIC, Australia
| | - Marcello G P Rosa
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia.,Australian Research Council, Centre of Excellence in Integrative Brain Function, Clayton, VIC, Australia
| | - Ramesh Rajan
- Biomedicine Discovery Institute and Department of Physiology, Monash University, Clayton, VIC, Australia.,Australian Research Council, Centre of Excellence in Integrative Brain Function, Clayton, VIC, Australia
| |
Collapse
|
23
|
Matsubara T, Ogata K, Hironaga N, Uehara T, Mitsudo T, Shigeto H, Maekawa T, Tobimatsu S. Monaural 40-Hz auditory steady-state magnetic responses can be useful for identifying epileptic focus in mesial temporal lobe epilepsy. Clin Neurophysiol 2018; 130:341-351. [PMID: 30669010 DOI: 10.1016/j.clinph.2018.11.026] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2018] [Revised: 10/19/2018] [Accepted: 11/28/2018] [Indexed: 02/03/2023]
Abstract
OBJECTIVE Patients with mesial temporal lobe epilepsy (mTLE) often exhibit central auditory processing (CAP) dysfunction. Monaural 40-Hz auditory steady-state magnetic responses (ASSRs) were recorded to explore the pathophysiology of mTLE. METHODS Eighteen left mTLE patients, 11 right mTLE patients and 16 healthy controls (HCs) were examined. Monaural clicks were presented at a rate of 40 Hz. Phase-locking factor (PLF) and power values were analyzed within bilateral Heschl's gyri. RESULTS Monaural 40-Hz ASSR demonstrated temporal frequency dynamics in both PLF and power data. Symmetrical hemispheric contralaterality was revealed in HCs. However, predominant contralaterality was absent in mTLE patients. Specifically, right mTLE patients exhibited a lack of contralaterality in response to left ear but not right ear stimulation, and vice versa in left mTLE patients. CONCLUSION This is the first study to use monaural 40-Hz ASSR with unilateral mTLE patients to clarify the relationship between CAP and epileptic focus. CAP dysfunction was characterized by a lack of contralaterality corresponding to epileptic focus. SIGNIFICANCE Monaural 40-Hz ASSR can provide useful information for localizing epileptic focus in mTLE patients.
Collapse
Affiliation(s)
- Teppei Matsubara
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan.
| | - Katsuya Ogata
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Naruhito Hironaga
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Taira Uehara
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Takako Mitsudo
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Hiroshi Shigeto
- Epilepsy and Sleep Center, Fukuoka Sanno Hospital, Fukuoka, Japan
| | | | - Shozo Tobimatsu
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| |
Collapse
|
24
|
Geissler DB, Weiler E, Ehret G. Adaptation and spectral enhancement at auditory temporal perceptual boundaries - Measurements via temporal precision of auditory brainstem responses. PLoS One 2018; 13:e0208935. [PMID: 30571726 PMCID: PMC6301773 DOI: 10.1371/journal.pone.0208935] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2018] [Accepted: 11/26/2018] [Indexed: 12/21/2022] Open
Abstract
In human and animal auditory perception the perceived quality of sound streams changes depending on the duration of inter-sound intervals (ISIs). Here, we studied whether adaptation and the precision of temporal coding in the auditory periphery reproduce general perceptual boundaries in the time domain near 20, 100, and 400 ms ISIs, the physiological origin of which are unknown. In four experiments, we recorded auditory brainstem responses with five wave peaks (P1 –P5) in response to acoustic models of communication calls of house mice, who perceived these calls with the mentioned boundaries. The newly introduced measure of average standard deviations of wave latencies of individual animals indicate the waves’ temporal precision (latency jitter) mostly in the range of 30–100 μs, very similar to latency jitter of single neurons. Adaptation effects of response latencies and latency jitter were measured for ISIs of 10–1000 ms. Adaptation decreased with increasing ISI duration following exponential or linear (on a logarithmic scale) functions in the range of up to about 200 ms ISIs. Adaptation effects were specific for each processing level in the auditory system. The perceptual boundaries near 20–30 and 100 ms ISIs were reflected in significant adaptation of latencies together with increases of latency jitter at P2-P5 for ISIs < ~30 ms and at P5 for ISIs < ~100 ms, respectively. Adaptation effects occurred when frequencies in a sound stream were within the same critical band. Ongoing low-frequency components/formants in a sound enhanced (decrease of latencies) coding of high-frequency components/formants when the frequencies concerned different critical bands. The results are discussed in the context of coding multi-harmonic sounds and stop-consonants-vowel pairs in the auditory brainstem. Furthermore, latency data at P1 (cochlea level) offer a reasonable value for the base-to-apex cochlear travel time in the mouse (0.342 ms) that has not been determined experimentally.
Collapse
Affiliation(s)
| | - Elke Weiler
- Institute of Neurobiology, University of Ulm, Ulm, Germany
| | - Günter Ehret
- Institute of Neurobiology, University of Ulm, Ulm, Germany
- * E-mail:
| |
Collapse
|
25
|
Matsubara T, Ogata K, Hironaga N, Kikuchi Y, Uehara T, Chatani H, Mitsudo T, Shigeto H, Tobimatsu S. Altered neural synchronization to pure tone stimulation in patients with mesial temporal lobe epilepsy: An MEG study. Epilepsy Behav 2018; 88:96-105. [PMID: 30243112 DOI: 10.1016/j.yebeh.2018.08.036] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/09/2018] [Revised: 08/28/2018] [Accepted: 08/29/2018] [Indexed: 11/25/2022]
Abstract
OBJECTIVE Our previous study of monaural auditory evoked magnetic fields (AEFs) demonstrated that hippocampal sclerosis significantly modulated auditory processing in patients with mesial temporal lobe epilepsy (mTLE). However, the small sample size (n = 17) and focus on the M100 response were insufficient to elucidate the lateralization of the epileptic focus. Therefore, we increased the number of patients with mTLE (n = 39) to examine whether neural synchronization induced by monaural pure tone stimulation provides useful diagnostic information about epileptic foci in patients with unilateral mTLE. METHODS Twenty-five patients with left mTLE, 14 patients with right mTLE, and 32 healthy controls (HCs) were recruited. Auditory stimuli of 500-Hz tone burst were monaurally presented to subjects. The AEF data were analyzed with source estimation of M100 responses in bilateral auditory cortices (ACs). Neural synchronization within ACs and between ACs was evaluated with phase-locking factor (PLF) and phase-locking value (PLV), respectively. Linear discriminant analysis was performed for diagnosis and lateralization of epileptic focus. RESULTS The M100 amplitude revealed that patients with right mTLE exhibited smaller M100 amplitude than patients with left mTLE and HCs. Interestingly, PLF was able to differentiate the groups with mTLE, with decreased PLFs in the alpha band observed in patients with right mTLE compared with those (PLFs) in patients with left mTLE. Right hemispheric predominance was confirmed in both HCs and patients with left mTLE while patients with right mTLE showed a lack of right hemispheric predominance. Functional connectivity between bilateral ACs (PLV) was reduced in both patients with right and left mTLE compared with that of HCs. The accuracy of diagnosis and lateralization was 80%-90%. CONCLUSION Auditory cortex subnormal function was more pronounced in patients with right mTLE compared with that in patients with left mTLE as well as HCs. Monaural AEFs can be used to reveal the pathophysiology of mTLE. Overall, our results indicate that altered neural synchronization may provide useful information about possible functional deterioration in patients with unilateral mTLE.
Collapse
Affiliation(s)
- Teppei Matsubara
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan.
| | - Katsuya Ogata
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Naruhito Hironaga
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Yoshikazu Kikuchi
- Department of Otorhinolaryngology, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Fukuoka, Japan
| | - Taira Uehara
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Hiroshi Chatani
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Takako Mitsudo
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| | - Hiroshi Shigeto
- Epilepsy and Sleep Center, Fukuoka Sanno Hospital, Fukuoka, Japan
| | - Shozo Tobimatsu
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Japan
| |
Collapse
|
26
|
Liang C, Houston LM, Samy RN, Abedelrehim LMI, Zhang F. Cortical Processing of Frequency Changes Reflected by the Acoustic Change Complex in Adult Cochlear Implant Users. Audiol Neurootol 2018; 23:152-164. [PMID: 30300882 DOI: 10.1159/000492170] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2017] [Accepted: 07/16/2018] [Indexed: 11/19/2022] Open
Abstract
The purpose of this study was to examine neural substrates of frequency change detection in cochlear implant (CI) recipients using the acoustic change complex (ACC), a type of cortical auditory evoked potential elicited by acoustic changes in an ongoing stimulus. A psychoacoustic test and electroencephalographic recording were administered in 12 postlingually deafened adult CI users. The stimuli were pure tones containing different magnitudes of upward frequency changes. Results showed that the frequency change detection threshold (FCDT) was 3.79% in the CI users, with a large variability. The ACC N1' latency was significantly correlated with the FCDT and the clinically collected speech perception score. The results suggested that the ACC evoked by frequency changes can serve as a useful objective tool in assessing frequency change detection capability and predicting speech perception performance in CI users.
Collapse
Affiliation(s)
- Chun Liang
- Department of Communication Sciences and Disorders, University of Cincinnati, Cincinnati, Ohio, USA.,Shenzhen Maternity and Child Healthcare Hospital, Shenzhen, China
| | - Lisa M Houston
- Department of Otolaryngology, Head and Neck Surgery, University of Cincinnati, Cincinnati, Ohio, USA
| | - Ravi N Samy
- Department of Otolaryngology, Head and Neck Surgery, University of Cincinnati, Cincinnati, Ohio, USA
| | - Lamiaa Mohamed Ibrahim Abedelrehim
- Department of Communication Sciences and Disorders, University of Cincinnati, Cincinnati, Ohio, USA.,Audiology Department, Sohag Faculty of Medicine, Sohag University, Sohag, Egypt
| | - Fawen Zhang
- Department of Communication Sciences and Disorders, University of Cincinnati, Cincinnati, Ohio,
| |
Collapse
|
27
|
Mehta K, Kliewer J, Ihlefeld A. Quantifying Neuronal Information Flow in Response to Frequency and Intensity Changes in the Auditory Cortex. CONFERENCE RECORD. ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS 2018; 2018:1367-1371. [PMID: 31595139 PMCID: PMC6782062 DOI: 10.1109/acssc.2018.8645091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Studies increasingly show that behavioral relevance alters the population representation of sensory stimuli in the sensory cortices. However, the mechanisms underlying this behavior are incompletely understood. Here, we record neuronal responses in the auditory cortex while a highly trained, awake, normal-hearing gerbil listens passively to target tones of high versus low behavioral relevance. Using an information theoretic framework, we model the overall transmission chain from acoustic input stimulus to recorded cortical response as a communication channel. To quantify how much information core auditory cortex carries about high versus low relevance sound, we then compute the mutual information of the multi-unit neuronal responses. Results show that the output over the stimulus-to-response channel can be modeled as a Poisson mixture. We derive a closed-form fast approximation for the entropy of a mixture of univariate Poisson random variables. A purely rate-code based model reveals reduced information transfer for high relevance compared to low relevance tones, hinting that changes in temporal discharge pattern may encode behavioral relevance.
Collapse
Affiliation(s)
- Ketan Mehta
- Krasnow Institute for Advanced Study, George Mason University, Fairfax, VA 22030
| | - Jörg Kliewer
- Helen and John C. Hartmann Dept. of Electrical and Computer Engineering New Jersey Institute of Technology, Newark, NJ 07102
| | - Antje Ihlefeld
- Dept. of Biomedical Engineering, New Jersey Institute of Technology, Newark, NJ 07102
| |
Collapse
|
28
|
Uluç I, Schmidt TT, Wu YH, Blankenburg F. Content-specific codes of parametric auditory working memory in humans. Neuroimage 2018; 183:254-262. [PMID: 30107259 DOI: 10.1016/j.neuroimage.2018.08.024] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2018] [Revised: 08/09/2018] [Accepted: 08/11/2018] [Indexed: 10/28/2022] Open
Abstract
Brain activity in frontal regions has been found to represent frequency information with a parametric code during working memory delay phases. The mental representation of frequencies has furthermore been shown to be modality independent in non-human primate electrophysiology and human EEG studies, suggesting frontal regions encoding quantitative information in a supramodal manner. A recent fMRI study using multivariate pattern analysis (MVPA) supports an overlapping multimodal network for the maintenance of visual and tactile frequency information over frontal and parietal brain regions. The present study extends the investigation of working memory representation of frequency information to the auditory domain. To this aim, we used MVPA on fMRI data recorded during an auditory frequency maintenance task. A support vector regression analysis revealed working memory information in auditory association areas and, consistent with earlier findings of parametric working memory, in a frontoparietal network. A direct comparison to an analogous dataset of vibrotactile parametric working memory revealed an overlap of information coding in prefrontal regions, particularly in the right inferior frontal gyrus. Therefore, our findings indicate that the prefrontal cortex represents frequency-specific working memory content irrespective of the modality as has been now also revealed for the auditory modality.
Collapse
Affiliation(s)
- Işıl Uluç
- Neurocomputation and Neuroimaging Unit (NNU), Department of Education and Psychology, Freie Universität Berlin, 14195 Berlin, Germany; Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, 10099 Berlin, Germany.
| | - Timo Torsten Schmidt
- Neurocomputation and Neuroimaging Unit (NNU), Department of Education and Psychology, Freie Universität Berlin, 14195 Berlin, Germany; Institute of Cognitive Science, University of Osnabrück, 49090 Osnabrück, Germany
| | - Yuan-Hao Wu
- Neurocomputation and Neuroimaging Unit (NNU), Department of Education and Psychology, Freie Universität Berlin, 14195 Berlin, Germany; Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, 10099 Berlin, Germany
| | - Felix Blankenburg
- Neurocomputation and Neuroimaging Unit (NNU), Department of Education and Psychology, Freie Universität Berlin, 14195 Berlin, Germany; Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, 10099 Berlin, Germany
| |
Collapse
|
29
|
Auditory priming improves neural synchronization in auditory-motor entrainment. Neuropsychologia 2018; 117:102-112. [DOI: 10.1016/j.neuropsychologia.2018.05.017] [Citation(s) in RCA: 26] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2017] [Revised: 05/18/2018] [Accepted: 05/20/2018] [Indexed: 11/18/2022]
|
30
|
Hamilton LS, Edwards E, Chang EF. A Spatial Map of Onset and Sustained Responses to Speech in the Human Superior Temporal Gyrus. Curr Biol 2018; 28:1860-1871.e4. [DOI: 10.1016/j.cub.2018.04.033] [Citation(s) in RCA: 98] [Impact Index Per Article: 16.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2017] [Revised: 03/04/2018] [Accepted: 04/10/2018] [Indexed: 01/05/2023]
|
31
|
Higgins I, Stringer S, Schnupp J. A Computational Account of the Role of Cochlear Nucleus and Inferior Colliculus in Stabilizing Auditory Nerve Firing for Auditory Category Learning. Neural Comput 2018; 30:1801-1829. [PMID: 29652586 DOI: 10.1162/neco_a_01085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
It is well known that auditory nerve (AN) fibers overcome bandwidth limitations through the volley principle, a form of multiplexing. What is less well known is that the volley principle introduces a degree of unpredictability into AN neural firing patterns that may be affecting even simple stimulus categorization learning. We use a physiologically grounded, unsupervised spiking neural network model of the auditory brain with spike time dependent plasticity learning to demonstrate that plastic auditory cortex is unable to learn even simple auditory object categories when exposed to the raw AN firing input without subcortical preprocessing. We then demonstrate the importance of nonplastic subcortical preprocessing within the cochlear nucleus and the inferior colliculus for stabilizing and denoising AN responses. Such preprocessing enables the plastic auditory cortex to learn efficient robust representations of the auditory object categories. The biological realism of our model makes it suitable for generating neurophysiologically testable hypotheses.
Collapse
Affiliation(s)
- Irina Higgins
- Department of Experimental Psychology, University of Oxford, Oxford, OX2 6GG, U.K.
| | - Simon Stringer
- Department of Experimental Psychology, University of Oxford, Oxford, OX2 6GG, U.K.
| | - Jan Schnupp
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, OX1 3QX, U.K.
| |
Collapse
|
32
|
Beebe NL, Schofield BR. Perineuronal nets in subcortical auditory nuclei of four rodent species with differing hearing ranges. J Comp Neurol 2018; 526:972-989. [PMID: 29277975 DOI: 10.1002/cne.24383] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2017] [Revised: 11/20/2017] [Accepted: 12/09/2017] [Indexed: 12/19/2022]
Abstract
Perineuronal nets (PNs) are aggregates of extracellular matrix molecules that surround some neurons in the brain. While PNs occur widely across many cortical areas, subcortical PNs are especially associated with motor and auditory systems. The auditory system has recently been suggested as an ideal model system for studying PNs and their functions. However, descriptions of PNs in subcortical auditory areas vary, and it is unclear whether the variation reflects species differences or differences in staining techniques. Here, we used two staining techniques (one lectin stain and one antibody stain) to examine PN distribution in the subcortical auditory system of four different species: guinea pigs (Cavia porcellus), mice (Mus musculus, CBA/CaJ strain), Long-Evans rats (Rattus norvegicus), and naked mole-rats (Heterocephalus glaber). We found that some auditory nuclei exhibit dramatic differences in PN distribution among species while other nuclei have consistent PN distributions. We also found that PNs exhibit molecular heterogeneity, and can stain with either marker individually or with both. PNs within a given nucleus can be heterogeneous or homogenous in their staining patterns. We compared PN staining across the frequency axes of tonotopically organized nuclei and among species with different hearing ranges. PNs were distributed non-uniformly across some nuclei, but only rarely did this appear related to the tonotopic axis. PNs were prominent in all four species; we found no systematic relationship between the hearing range and the number, staining patterns or distribution of PNs in the auditory nuclei.
Collapse
Affiliation(s)
- Nichole L Beebe
- Department of Anatomy and Neurobiology, Northeast Ohio Medical University, Rootstown, Ohio, 44272
| | - Brett R Schofield
- Department of Anatomy and Neurobiology, Northeast Ohio Medical University, Rootstown, Ohio, 44272
| |
Collapse
|
33
|
|
34
|
Nozaradan S, Keller PE, Rossion B, Mouraux A. EEG Frequency-Tagging and Input-Output Comparison in Rhythm Perception. Brain Topogr 2017; 31:153-160. [PMID: 29127530 DOI: 10.1007/s10548-017-0605-8] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2017] [Accepted: 10/27/2017] [Indexed: 01/23/2023]
Abstract
The combination of frequency-tagging with electroencephalography (EEG) has recently proved fruitful for understanding the perception of beat and meter in musical rhythm, a common behavior shared by humans of all cultures. EEG frequency-tagging allows the objective measurement of input-output transforms to investigate beat perception, its modulation by exogenous and endogenous factors, development, and neural basis. Recent doubt has been raised about the validity of comparing frequency-domain representations of auditory rhythmic stimuli and corresponding EEG responses, assuming that it implies a one-to-one mapping between the envelope of the rhythmic input and the neural output, and that it neglects the sensitivity of frequency-domain representations to acoustic features making up the rhythms. Here we argue that these elements actually reinforce the strengths of the approach. The obvious fact that acoustic features influence the frequency spectrum of the sound envelope precisely justifies taking into consideration the sounds used to generate a beat percept for interpreting neural responses to auditory rhythms. Most importantly, the many-to-one relationship between rhythmic input and perceived beat actually validates an approach that objectively measures the input-output transforms underlying the perceptual categorization of rhythmic inputs. Hence, provided that a number of potential pitfalls and fallacies are avoided, EEG frequency-tagging to study input-output relationships appears valuable for understanding rhythm perception.
Collapse
Affiliation(s)
- Sylvie Nozaradan
- The MARCS Institute for Brain, Behaviour and Development (WSU), Sydney, NSW, Australia. .,Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), Brussels, Belgium. .,International Laboratory for Brain, Music and Sound Research (Brams), Montreal, QC, Canada. .,MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Locked Bag 1797, Penrith, NSW, 2751, Australia.
| | - Peter E Keller
- The MARCS Institute for Brain, Behaviour and Development (WSU), Sydney, NSW, Australia
| | - Bruno Rossion
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), Brussels, Belgium.,Neurology Unit, Centre Hospitalier Régional Universitaire (CHRU) de Nancy, Nancy, France
| | - André Mouraux
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), Brussels, Belgium
| |
Collapse
|
35
|
Nozaradan S, Schwartze M, Obermeier C, Kotz SA. Specific contributions of basal ganglia and cerebellum to the neural tracking of rhythm. Cortex 2017; 95:156-168. [DOI: 10.1016/j.cortex.2017.08.015] [Citation(s) in RCA: 60] [Impact Index Per Article: 8.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2017] [Revised: 07/16/2017] [Accepted: 08/07/2017] [Indexed: 11/29/2022]
|
36
|
Holdgraf CR, Rieger JW, Micheli C, Martin S, Knight RT, Theunissen FE. Encoding and Decoding Models in Cognitive Electrophysiology. Front Syst Neurosci 2017; 11:61. [PMID: 29018336 PMCID: PMC5623038 DOI: 10.3389/fnsys.2017.00061] [Citation(s) in RCA: 64] [Impact Index Per Article: 9.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2017] [Accepted: 08/07/2017] [Indexed: 11/13/2022] Open
Abstract
Cognitive neuroscience has seen rapid growth in the size and complexity of data recorded from the human brain as well as in the computational tools available to analyze this data. This data explosion has resulted in an increased use of multivariate, model-based methods for asking neuroscience questions, allowing scientists to investigate multiple hypotheses with a single dataset, to use complex, time-varying stimuli, and to study the human brain under more naturalistic conditions. These tools come in the form of "Encoding" models, in which stimulus features are used to model brain activity, and "Decoding" models, in which neural features are used to generated a stimulus output. Here we review the current state of encoding and decoding models in cognitive electrophysiology and provide a practical guide toward conducting experiments and analyses in this emerging field. Our examples focus on using linear models in the study of human language and audition. We show how to calculate auditory receptive fields from natural sounds as well as how to decode neural recordings to predict speech. The paper aims to be a useful tutorial to these approaches, and a practical introduction to using machine learning and applied statistics to build models of neural activity. The data analytic approaches we discuss may also be applied to other sensory modalities, motor systems, and cognitive systems, and we cover some examples in these areas. In addition, a collection of Jupyter notebooks is publicly available as a complement to the material covered in this paper, providing code examples and tutorials for predictive modeling in python. The aim is to provide a practical understanding of predictive modeling of human brain data and to propose best-practices in conducting these analyses.
Collapse
Affiliation(s)
- Christopher R. Holdgraf
- Department of Psychology, Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, United States
- Office of the Vice Chancellor for Research, Berkeley Institute for Data Science, University of California, Berkeley, Berkeley, CA, United States
| | - Jochem W. Rieger
- Department of Psychology, Carl-von-Ossietzky University, Oldenburg, Germany
| | - Cristiano Micheli
- Department of Psychology, Carl-von-Ossietzky University, Oldenburg, Germany
- Institut des Sciences Cognitives Marc Jeannerod, Lyon, France
| | - Stephanie Martin
- Department of Psychology, Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, United States
- Defitech Chair in Brain-Machine Interface, Center for Neuroprosthetics, Ecole Polytechnique Fédérale de Lausanne, Lausanne, Switzerland
| | - Robert T. Knight
- Department of Psychology, Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, United States
| | - Frederic E. Theunissen
- Department of Psychology, Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA, United States
- Department of Psychology, University of California, Berkeley, Berkeley, CA, United States
| |
Collapse
|
37
|
Higgins I, Stringer S, Schnupp J. Unsupervised learning of temporal features for word categorization in a spiking neural network model of the auditory brain. PLoS One 2017; 12:e0180174. [PMID: 28797034 PMCID: PMC5552261 DOI: 10.1371/journal.pone.0180174] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2016] [Accepted: 06/12/2017] [Indexed: 12/04/2022] Open
Abstract
The nature of the code used in the auditory cortex to represent complex auditory stimuli, such as naturally spoken words, remains a matter of debate. Here we argue that such representations are encoded by stable spatio-temporal patterns of firing within cell assemblies known as polychronous groups, or PGs. We develop a physiologically grounded, unsupervised spiking neural network model of the auditory brain with local, biologically realistic, spike-time dependent plasticity (STDP) learning, and show that the plastic cortical layers of the network develop PGs which convey substantially more information about the speaker independent identity of two naturally spoken word stimuli than does rate encoding that ignores the precise spike timings. We furthermore demonstrate that such informative PGs can only develop if the input spatio-temporal spike patterns to the plastic cortical areas of the model are relatively stable.
Collapse
Affiliation(s)
- Irina Higgins
- Department of Experimental Psychology, University of Oxford, Oxford, England
| | - Simon Stringer
- Department of Experimental Psychology, University of Oxford, Oxford, England
| | - Jan Schnupp
- Department of Physiology, Anatomy and Genetics (DPAG), University of Oxford, Oxford, England
| |
Collapse
|
38
|
Smith NA, Folland NA, Martinez DM, Trainor LJ. Multisensory object perception in infancy: 4-month-olds perceive a mistuned harmonic as a separate auditory and visual object. Cognition 2017; 164:1-7. [PMID: 28346869 PMCID: PMC5429982 DOI: 10.1016/j.cognition.2017.01.016] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2015] [Revised: 01/17/2017] [Accepted: 01/24/2017] [Indexed: 10/19/2022]
Abstract
Infants learn to use auditory and visual information to organize the sensory world into identifiable objects with particular locations. Here we use a behavioural method to examine infants' use of harmonicity cues to auditory object perception in a multisensory context. Sounds emitted by different objects sum in the air and the auditory system must figure out which parts of the complex waveform belong to different sources (auditory objects). One important cue to this source separation is that complex tones with pitch typically contain a fundamental frequency and harmonics at integer multiples of the fundamental. Consequently, adults hear a mistuned harmonic in a complex sound as a distinct auditory object (Alain, Theunissen, Chevalier, Batty, & Taylor, 2003). Previous work by our group demonstrated that 4-month-old infants are also sensitive to this cue. They behaviourally discriminate a complex tone with a mistuned harmonic from the same complex with in-tune harmonics, and show an object-related event-related potential (ERP) electrophysiological (EEG) response to the stimulus with mistuned harmonics. In the present study we use an audiovisual procedure to investigate whether infants perceive a complex tone with an 8% mistuned harmonic as emanating from two objects, rather than merely detecting the mistuned cue. We paired in-tune and mistuned complex tones with visual displays that contained either one or two bouncing balls. Four-month-old infants showed surprise at the incongruous pairings, looking longer at the display of two balls when paired with the in-tune complex and at the display of one ball when paired with the mistuned harmonic complex. We conclude that infants use harmonicity as a cue for source separation when integrating auditory and visual information in object perception.
Collapse
Affiliation(s)
- Nicholas A Smith
- Perceptual Development Laboratory, Boys Town National Research Hospital, 555 N. 30th Street, Omaha, NE 68131, United States
| | - Nicole A Folland
- Department of Psychology, Neuroscience and Behaviour, McMaster University, 1280 Main Street West, Hamilton, Ontario L8S 4K1, Canada
| | - Diana M Martinez
- Department of Psychology, Neuroscience and Behaviour, McMaster University, 1280 Main Street West, Hamilton, Ontario L8S 4K1, Canada
| | - Laurel J Trainor
- Department of Psychology, Neuroscience and Behaviour, McMaster University, 1280 Main Street West, Hamilton, Ontario L8S 4K1, Canada; McMaster Institute for Music and the Mind, McMaster University, 1280 Main Street West, Hamilton, Ontario L8S 4K1, Canada; Rotman Research Institute, Baycrest, University of Toronto, 3560 Bathurst Street, Toronto, Ontario M6A 2E1, Canada.
| |
Collapse
|
39
|
Bellur A, Elhilali M. Feedback-Driven Sensory Mapping Adaptation for Robust Speech Activity Detection. IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING 2017; 25:481-492. [PMID: 28736736 PMCID: PMC5516649 DOI: 10.1109/taslp.2016.2639322] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Parsing natural acoustic scenes using computational methodologies poses many challenges. Given the rich and complex nature of the acoustic environment, data mismatch between train and test conditions is a major hurdle in data-driven audio processing systems. In contrast, the brain exhibits a remarkable ability at segmenting acoustic scenes with relative ease. When tackling challenging listening conditions that are often faced in everyday life, the biological system relies on a number of principles that allow it to effortlessly parse its rich soundscape. In the current study, we leverage a key principle employed by the auditory system: its ability to adapt the neural representation of its sensory input in a high-dimensional space. We propose a framework that mimics this process in a computational model for robust speech activity detection. The system employs a 2-D Gabor filter bank whose parameters are retuned offline to improve the separability between the feature representation of speech and nonspeech sounds. This retuning process, driven by feedback from statistical models of speech and nonspeech classes, attempts to minimize the misclassification risk of mismatched data, with respect to the original statistical models. We hypothesize that this risk minimization procedure results in an emphasis of unique speech and nonspeech modulations in the high-dimensional space. We show that such an adapted system is indeed robust to other novel conditions, with a marked reduction in equal error rates for a variety of databases with additive and convolutive noise distortions. We discuss the lessons learned from biology with regard to adapting to an ever-changing acoustic environment and the impact on building truly intelligent audio processing systems.
Collapse
Affiliation(s)
- Ashwin Bellur
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218 USA
| | - Mounya Elhilali
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218 USA
| |
Collapse
|
40
|
Kikuchi Y, Okamoto T, Ogata K, Hagiwara K, Umezaki T, Kenjo M, Nakagawa T, Tobimatsu S. Abnormal auditory synchronization in stuttering: A magnetoencephalographic study. Hear Res 2017; 344:82-89. [DOI: 10.1016/j.heares.2016.10.027] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/29/2016] [Accepted: 10/31/2016] [Indexed: 02/01/2023]
|
41
|
Holdgraf CR, de Heer W, Pasley B, Rieger J, Crone N, Lin JJ, Knight RT, Theunissen FE. Rapid tuning shifts in human auditory cortex enhance speech intelligibility. Nat Commun 2016; 7:13654. [PMID: 27996965 PMCID: PMC5187445 DOI: 10.1038/ncomms13654] [Citation(s) in RCA: 49] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2016] [Accepted: 10/20/2016] [Indexed: 11/11/2022] Open
Abstract
Experience shapes our perception of the world on a moment-to-moment basis. This robust perceptual effect of experience parallels a change in the neural representation of stimulus features, though the nature of this representation and its plasticity are not well-understood. Spectrotemporal receptive field (STRF) mapping describes the neural response to acoustic features, and has been used to study contextual effects on auditory receptive fields in animal models. We performed a STRF plasticity analysis on electrophysiological data from recordings obtained directly from the human auditory cortex. Here, we report rapid, automatic plasticity of the spectrotemporal response of recorded neural ensembles, driven by previous experience with acoustic and linguistic information, and with a neurophysiological effect in the sub-second range. This plasticity reflects increased sensitivity to spectrotemporal features, enhancing the extraction of more speech-like features from a degraded stimulus and providing the physiological basis for the observed ‘perceptual enhancement' in understanding speech. Experience constantly shapes perception, but the neural mechanisms of this rapid plasticity are unclear. Here, Holdgraf et al. record neural activity in the human auditory cortex and show that listening to normal speech elicits rapid plasticity that increases the neural gain for features of sound that are key for speech intelligibility.
Collapse
Affiliation(s)
- Christopher R Holdgraf
- Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720, USA
| | - Wendy de Heer
- Department of Psychology, University of California, Berkeley, California 94720, USA
| | - Brian Pasley
- Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720, USA
| | - Jochem Rieger
- Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720, USA
| | - Nathan Crone
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, Maryland 21205, USA
| | - Jack J Lin
- UC Irvine Comprehensive Epilepsy Program, Department of Neurology, University of California, Irvine, California 92868, USA
| | - Robert T Knight
- Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720, USA.,Department of Psychology, University of California, Berkeley, California 94720, USA.,Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, Maryland 21205, USA
| | - Frédéric E Theunissen
- Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720, USA.,Department of Psychology, University of California, Berkeley, California 94720, USA
| |
Collapse
|
42
|
Nozaradan S, Mouraux A, Jonas J, Colnat-Coulbois S, Rossion B, Maillard L. Intracerebral evidence of rhythm transform in the human auditory cortex. Brain Struct Funct 2016; 222:2389-2404. [PMID: 27990557 DOI: 10.1007/s00429-016-1348-0] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2016] [Accepted: 12/06/2016] [Indexed: 01/23/2023]
Abstract
Musical entrainment is shared by all human cultures and the perception of a periodic beat is a cornerstone of this entrainment behavior. Here, we investigated whether beat perception might have its roots in the earliest stages of auditory cortical processing. Local field potentials were recorded from 8 patients implanted with depth-electrodes in Heschl's gyrus and the planum temporale (55 recording sites in total), usually considered as human primary and secondary auditory cortices. Using a frequency-tagging approach, we show that both low-frequency (<30 Hz) and high-frequency (>30 Hz) neural activities in these structures faithfully track auditory rhythms through frequency-locking to the rhythm envelope. A selective gain in amplitude of the response frequency-locked to the beat frequency was observed for the low-frequency activities but not for the high-frequency activities, and was sharper in the planum temporale, especially for the more challenging syncopated rhythm. Hence, this gain process is not systematic in all activities produced in these areas and depends on the complexity of the rhythmic input. Moreover, this gain was disrupted when the rhythm was presented at fast speed, revealing low-pass response properties which could account for the propensity to perceive a beat only within the musical tempo range. Together, these observations show that, even though part of these neural transforms of rhythms could already take place in subcortical auditory processes, the earliest auditory cortical processes shape the neural representation of rhythmic inputs in favor of the emergence of a periodic beat.
Collapse
Affiliation(s)
- Sylvie Nozaradan
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), 53, Avenue Mounier, UCL 53.75, 1200, Brussels, Belgium. .,The MARCS Institute, Western Sydney University, Sydney, NSW, 2214, Australia. .,International Laboratory for Brain, Music and Sound Research (Brams), Montreal, H3C 3J7, Canada.
| | - André Mouraux
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), 53, Avenue Mounier, UCL 53.75, 1200, Brussels, Belgium
| | - Jacques Jonas
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), 53, Avenue Mounier, UCL 53.75, 1200, Brussels, Belgium.,Service de Neurologie, Centre Hospitalier Universitaire de Nancy, 54035, Nancy, France.,CRAN UMR 7039 CNRS Université de Lorraine, 54035, Nancy, France
| | - Sophie Colnat-Coulbois
- Neurosurgery Department, Centre Hospitalier Universitaire de Nancy, 54035, Nancy, France
| | - Bruno Rossion
- Institute of Neuroscience (Ions), Université catholique de Louvain (UCL), 53, Avenue Mounier, UCL 53.75, 1200, Brussels, Belgium.,Service de Neurologie, Centre Hospitalier Universitaire de Nancy, 54035, Nancy, France.,Psychological Sciences Research Institute, Université Catholique de Louvain (UCL), 1348, Louvain-la-Neuve, Belgium
| | - Louis Maillard
- Service de Neurologie, Centre Hospitalier Universitaire de Nancy, 54035, Nancy, France.,CRAN UMR 7039 CNRS Université de Lorraine, 54035, Nancy, France
| |
Collapse
|
43
|
Neural constraints and flexibility in language processing. Behav Brain Sci 2016; 39:e78. [PMID: 27561969 DOI: 10.1017/s0140525x15000837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Humans process language with their neurons. Memory in neurons is supported by neural firing and by short- and long-term synaptic weight change; the emergent behaviour of neurons, synchronous firing, and cell assembly dynamics is also a form of memory. As the language signal moves to later stages, it is processed with different mechanisms that are slower but more persistent.
Collapse
|
44
|
Detection efficiency of auditory steady state evoked by modulated noise. Hear Res 2016; 339:125-31. [PMID: 27262450 DOI: 10.1016/j.heares.2016.05.017] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/20/2015] [Revised: 05/23/2016] [Accepted: 05/26/2016] [Indexed: 11/21/2022]
Abstract
AIM This study aimed to investigate the efficiency of Magnitude Squared Coherence (MSC) and Spectral F test (SFT) for the detection of auditory steady state responses (ASSR) obtained by amplitude-modulated noises. MATERIAL AND METHODS Twenty individuals (12 women) without any history of neurological or audiological diseases, aged from 18 to 59 years (mean ± standard deviation = 26.45 ± 3.9 years), who provided written informed consent, participated in the study. The Audiostim system was used for stimulating and ASSR recording. The tested stimuli were amplitude-modulated Wide-band noise (WBN), Low-band noise (LBN), High-band noise (HBN), Two-band noise (TBN) between 77 and 110 Hz, applied in intensity levels of 55, 45, and 25 dB sound pressure level (SPL). MSC and SFT, two statistical-based detection techniques, were applied with a significance level of 5%. Detection times and rates were compared using the Friedman test and Tukey-Kramer as post hoc analysis. Also based on the stimulation parameters (stimuli types and intensity levels) and detection techniques (MSC or SFT), 16 different pass/fail protocols, for which the true negatives (TN) were calculated. RESULTS The median detection times ranged from 68 to 157s for 55 dB SPL, 68-99s for 45 dB SPL, and 84-118s for 25 dB SPL. No statistical difference was found between MSC and STF considering the median detection times (p > 0.05). The detection rates ranged from 100% to 55.6% in 55 dB SPL, 97.2%-38.9% in 45 dB SPL and 66.7%-8.3% in 25 dB SPL. Also for detection rates, no statistical difference was observed between MSC and STF (p > 0.05). True negatives (TN) above 90% were found for Protocols that employed WBN or HBN, at 55 dB SPL or that used WBN or HBN, at 45 dB SPL. For Protocols employing TBN, at 55 dB SPL or 45 dB SPL TN below 60% were found due to the low detection rates of stimuli that included low-band frequencies. CONCLUSION The stimuli that include high-frequency content showed higher detection rates (>90%) and lower detection times (<3 min). The noise composed by two bands applied separately (TBN) is not feasible for clinical applications since it requires prolonging the exam duration, and also led to a reduced percentage of true negatives. On the other hand, WBN and HBN achieved high detection performance and high TN and should be investigated to implement pass/fail protocol for hearing screening with clinical population. Finally, both WBN and HBN seemed to be indifferent to the employed technique (SFT or MSC), which can be seen as another advantage of ASSR employment.
Collapse
|
45
|
O'Brien GE, Imennov NS, Rubinstein JT. Simulating electrical modulation detection thresholds using a biophysical model of the auditory nerve. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:2448. [PMID: 27250141 DOI: 10.1121/1.4947430] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Modulation detection thresholds (MDTs) assess listeners' sensitivity to changes in the temporal envelope of a signal and have been shown to strongly correlate with speech perception in cochlear implant users. MDTs are simulated with a stochastic model of a population of auditory nerve fibers that has been verified to accurately simulate a number of physiologically important temporal response properties. The procedure to estimate detection thresholds has previously been applied to stimulus discrimination tasks. The population model simulates the MDT-stimulus intensity relationship measured in cochlear implant users. The model also recreates the shape of the modulation transfer function and the relationship between MDTs and carrier rate. Discrimination based on fluctuations in synchronous firing activity predicts better performance at low carrier rates, but quantitative measures of modulation coding predict better neural representation of high carrier rate stimuli. Manipulating the number of fibers and a temporal integration parameter, the width of a sliding temporal integration window, varies properties of the MDTs, such as cutoff frequency and peak threshold. These results demonstrate the importance of using a multi-diameter fiber population in modeling the MDTs and demonstrate a wider applicability of this model to simulating behavioral performance in cochlear implant listeners.
Collapse
Affiliation(s)
- Gabrielle E O'Brien
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| | - Nikita S Imennov
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| | - Jay T Rubinstein
- Department of Otolaryngology, V. M. Bloedel Hearing Research Center, University of Washington, Box 3657923, CHDD building, CD 176, Seattle, Washington 98196, USA
| |
Collapse
|
46
|
Tang H, Crain S, Johnson BW. Dual temporal encoding mechanisms in human auditory cortex: Evidence from MEG and EEG. Neuroimage 2016; 128:32-43. [DOI: 10.1016/j.neuroimage.2015.12.053] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2015] [Revised: 12/01/2015] [Accepted: 12/30/2015] [Indexed: 11/25/2022] Open
|
47
|
Jain S, Dwarkanath VM. Effect of tinnitus location on the psychoacoustic measures of hearing. HEARING BALANCE AND COMMUNICATION 2015. [DOI: 10.3109/21695717.2016.1099885] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
|
48
|
Pannese A, Grandjean D, Frühholz S. Subcortical processing in auditory communication. Hear Res 2015; 328:67-77. [DOI: 10.1016/j.heares.2015.07.003] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/03/2015] [Revised: 06/23/2015] [Accepted: 07/01/2015] [Indexed: 12/21/2022]
|
49
|
Almeqbel A, McMahon C. Objective measurement of high-level auditory cortical function in children. Int J Pediatr Otorhinolaryngol 2015; 79:1055-62. [PMID: 25998216 DOI: 10.1016/j.ijporl.2015.04.026] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/18/2015] [Revised: 04/16/2015] [Accepted: 04/17/2015] [Indexed: 10/23/2022]
Abstract
OBJECTIVE This study examined whether the N2 latency of the cortical auditory evoked potential (CAEP) could be used as an objective indicator of temporal processing ability in normally hearing children. METHODS The N2 latency was evoked using three temporal processing paradigms: (1) differences in voice-onset-times (VOTs); (2) speech-in-noise using the CV/da/embedded in broadband noise (BBN) with varying signal-to-noise ratios (SNRs); and (3) 16Hz amplitude-modulated (AM) BBN presented (i) alone and (ii) following an unmodulated BBN, using four modulation depths. Thirty-four school-aged children with normal hearing, speech, language and reading were stratified into two groups: 5-7 years (n=13) and 8-12 years (n=21). RESULTS The N2 latency shifted significantly and systematically with differences in VOT and SNR, and was significantly different in the two AM-BBN conditions. CONCLUSIONS For children without an N1 peak in the cortical waveform, the N2 peak can be used as a sensitive measure of temporal processing for these stimuli. SIGNIFICANCE N2 latency of the CAEP can be used as an objective measure of temporal processing ability in a paediatric population with temporal processing disorder who are difficult to assess via behavioural response.
Collapse
Affiliation(s)
- Aseel Almeqbel
- Department of Hearing and Speech Sciences, Faculty of Allied Health Sciences, Health Sciences Center, Kuwait University, Kuwait City, Kuwait.
| | - Catherine McMahon
- Linguistics Department, Faculty of Human Sciences, Macquarie University, Sydney, NSW, Australia; The HEARing Cooperative Research Centre (CRC), Melbourne, VIC, Australia
| |
Collapse
|
50
|
Berger JI, Coomber B. Tinnitus-related changes in the inferior colliculus. Front Neurol 2015; 6:61. [PMID: 25870582 PMCID: PMC4378364 DOI: 10.3389/fneur.2015.00061] [Citation(s) in RCA: 29] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2014] [Accepted: 03/09/2015] [Indexed: 12/21/2022] Open
Abstract
Tinnitus is highly complex, diverse, and difficult to treat, in part due to the fact that the underlying causes and mechanisms remain elusive. Tinnitus is generated within the auditory brain; however, consolidating our understanding of tinnitus pathophysiology is difficult due to the diversity of reported effects and the variety of implicated brain nuclei. Here, we focus on the inferior colliculus (IC), a midbrain structure that integrates the vast majority of ascending auditory information and projects via the thalamus to the auditory cortex. The IC is also a point of convergence for corticofugal input and input originating outside the auditory pathway. We review the evidence, from both studies with human subjects and from animal models, for the contribution the IC makes to tinnitus. Changes in the IC, caused by either noise exposure or drug administration, involve fundamental, heterogeneous alterations in the balance of excitation and inhibition. However, differences between hearing loss-induced pathology and tinnitus-related pathology are not well understood. Moreover, variability in tinnitus induction methodology has a significant impact on subsequent neural and behavioral changes, which could explain some of the seemingly contradictory data. Nonetheless, the IC is likely involved in the generation and persistence of tinnitus perception.
Collapse
Affiliation(s)
- Joel I Berger
- Medical Research Council Institute of Hearing Research, University of Nottingham , Nottingham , UK
| | - Ben Coomber
- Medical Research Council Institute of Hearing Research, University of Nottingham , Nottingham , UK
| |
Collapse
|