1
|
Kachlicka M, Tierney A. Voice actors show enhanced neural tracking of pitch, prosody perception, and music perception. Cortex 2024; 178:213-222. [PMID: 39024939 DOI: 10.1016/j.cortex.2024.06.016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 05/28/2024] [Accepted: 06/26/2024] [Indexed: 07/20/2024]
Abstract
Experiences with sound that make strong demands on the precision of perception, such as musical training and experience speaking a tone language, can enhance auditory neural encoding. Are high demands on the precision of perception necessary for training to drive auditory neural plasticity? Voice actors are an ideal subject population for answering this question. Voice acting requires exaggerating prosodic cues to convey emotion, character, and linguistic structure, drawing upon attention to sound, memory for sound features, and accurate sound production, but not fine perceptual precision. Here we assessed neural encoding of pitch using the frequency-following response (FFR), as well as prosody, music, and sound perception, in voice actors and a matched group of non-actors. We find that the consistency of neural sound encoding, prosody perception, and musical phrase perception are all enhanced in voice actors, suggesting that a range of neural and behavioural auditory processing enhancements can result from training which lacks fine perceptual precision. However, fine discrimination was not enhanced in voice actors but was linked to degree of musical experience, suggesting that low-level auditory processing can only be enhanced by demanding perceptual training. These findings suggest that training which taxes attention, memory, and production but is not perceptually taxing may be a way to boost neural encoding of sound and auditory pattern detection in individuals with poor auditory skills.
Collapse
Affiliation(s)
- Magdalena Kachlicka
- School of Psychological Sciences, Birkbeck, University of London, London, UK
| | - Adam Tierney
- School of Psychological Sciences, Birkbeck, University of London, London, UK.
| |
Collapse
|
2
|
Han M, Chien YF, Zhang Z, Wei Z, Li W. Music training affects listeners' processing of different types of accentuation information: Evidence from ERPs. Brain Cogn 2024; 174:106120. [PMID: 38142535 DOI: 10.1016/j.bandc.2023.106120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2023] [Revised: 12/07/2023] [Accepted: 12/14/2023] [Indexed: 12/26/2023]
Abstract
Previous studies found that prolonged musical training can promote language processing, but few studies have examined whether and how musical training affects the processing of accentuation in spoken language. In this study, a vocabulary detection task was conducted, with Chinese single sentences as materials, to investigate how musicians and non-musicians process corrective accent and information accent in the sentence-middle and sentence-final positions. In the sentence-middle position, results of the cluster-based permutation t-tests showed significant differences in the 574-714 ms time window for the control group. In the sentence-final position, the cluster-based permutation t-tests revealed significant differences in the 612-810 ms time window for the music group and in the 616-812 ms time window for the control group. These significant positive effects were induced by the processing of information accent relative to that of corrective accent. These results suggest that both groups were able to distinguish corrective accent from information accent, but they processed the two accent types differently in the sentence-middle position. These findings show that musical training has a cross-domain effect on spoken language processing and that the accent position also affects its processing.
Collapse
Affiliation(s)
- Mei Han
- School of Public Health, Bengbu Medical University, China; Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, China; Key Laboratory of Brain and Cognitive Neuroscience, Liaoning Province, China
| | - Yu-Fu Chien
- Department of Chinese Language and Literature, Fudan University, China
| | - Zhenghua Zhang
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, China; Key Laboratory of Brain and Cognitive Neuroscience, Liaoning Province, China; Department of Psychology, Renmin University of China, Beijing, China
| | - Zhen Wei
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, China; Key Laboratory of Brain and Cognitive Neuroscience, Liaoning Province, China
| | - Weijun Li
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, China; Key Laboratory of Brain and Cognitive Neuroscience, Liaoning Province, China.
| |
Collapse
|
3
|
Rizzi R, Bidelman GM. Duplex perception reveals brainstem auditory representations are modulated by listeners' ongoing percept for speech. Cereb Cortex 2023; 33:10076-10086. [PMID: 37522248 PMCID: PMC10502779 DOI: 10.1093/cercor/bhad266] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 06/27/2023] [Accepted: 07/10/2023] [Indexed: 08/01/2023] Open
Abstract
So-called duplex speech stimuli with perceptually ambiguous spectral cues to one ear and isolated low- versus high-frequency third formant "chirp" to the opposite ear yield a coherent percept supporting their phonetic categorization. Critically, such dichotic sounds are only perceived categorically upon binaural integration. Here, we used frequency-following responses (FFRs), scalp-recorded potentials reflecting phase-locked subcortical activity, to investigate brainstem responses to fused speech percepts and to determine whether FFRs reflect binaurally integrated category-level representations. We recorded FFRs to diotic and dichotic stop-consonants (/da/, /ga/) that either did or did not require binaural fusion to properly label along with perceptually ambiguous sounds without clear phonetic identity. Behaviorally, listeners showed clear categorization of dichotic speech tokens confirming they were heard with a fused, phonetic percept. Neurally, we found FFRs were stronger for categorically perceived speech relative to category-ambiguous tokens but also differentiated phonetic categories for both diotically and dichotically presented speech sounds. Correlations between neural and behavioral data further showed FFR latency predicted the degree to which listeners labeled tokens as "da" versus "ga." The presence of binaurally integrated, category-level information in FFRs suggests human brainstem processing reflects a surprisingly abstract level of the speech code typically circumscribed to much later cortical processing.
Collapse
Affiliation(s)
- Rose Rizzi
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
- Program in Neuroscience, Indiana University, Bloomington, IN, United States
- School of Communication Sciences and Disorders, University of Memphis, Memphis, TN, United States
| | - Gavin M Bidelman
- Department of Speech, Language, and Hearing Sciences, Indiana University, Bloomington, IN, United States
- Program in Neuroscience, Indiana University, Bloomington, IN, United States
- Cognitive Science Program, Indiana University, Bloomington, IN, United States
| |
Collapse
|
4
|
Willmore BDB, King AJ. Adaptation in auditory processing. Physiol Rev 2023; 103:1025-1058. [PMID: 36049112 PMCID: PMC9829473 DOI: 10.1152/physrev.00011.2022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/21/2023] Open
Abstract
Adaptation is an essential feature of auditory neurons, which reduces their responses to unchanging and recurring sounds and allows their response properties to be matched to the constantly changing statistics of sounds that reach the ears. As a consequence, processing in the auditory system highlights novel or unpredictable sounds and produces an efficient representation of the vast range of sounds that animals can perceive by continually adjusting the sensitivity and, to a lesser extent, the tuning properties of neurons to the most commonly encountered stimulus values. Together with attentional modulation, adaptation to sound statistics also helps to generate neural representations of sound that are tolerant to background noise and therefore plays a vital role in auditory scene analysis. In this review, we consider the diverse forms of adaptation that are found in the auditory system in terms of the processing levels at which they arise, the underlying neural mechanisms, and their impact on neural coding and perception. We also ask what the dynamics of adaptation, which can occur over multiple timescales, reveal about the statistical properties of the environment. Finally, we examine how adaptation to sound statistics is influenced by learning and experience and changes as a result of aging and hearing loss.
Collapse
Affiliation(s)
- Ben D. B. Willmore
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| | - Andrew J. King
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
5
|
Easwar V, Chung L. The influence of phoneme contexts on adaptation in vowel-evoked envelope following responses. Eur J Neurosci 2022; 56:4572-4582. [PMID: 35804282 PMCID: PMC9543495 DOI: 10.1111/ejn.15768] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2021] [Revised: 02/25/2022] [Accepted: 07/06/2022] [Indexed: 11/28/2022]
Abstract
Repeated stimulus presentation leads to neural adaptation and consequent amplitude reduction in vowel-evoked envelope following responses (EFRs)-a response that reflects neural activity phase-locked to envelope periodicity. EFRs are elicited by vowels presented in isolation or in the context of other phonemes such as in syllables. While context phonemes could exert some forward influence on vowel-evoked EFRs, they may reduce the degree of adaptation. Here, we evaluated whether the properties of context phonemes between consecutive vowel stimuli influence adaptation. EFRs were elicited by the low-frequency first formant (resolved harmonics) and mid-to-high frequency second and higher formants (unresolved harmonics) of a male-spoken/i/when the presence, number, and predictability of context phonemes (/s/, /a/, /∫/, /u/) between vowel repetitions varied. Monitored over four iterations of /i/, adaptation was evident only for EFRs elicited by the unresolved harmonics. EFRs elicited by the unresolved harmonics decreased in amplitude by ~16-20 nV (10-17%) after the first presentation of/i/and remained stable thereafter. EFR adaptation was reduced by the presence of a context phoneme, but the reduction did not change with their number or predictability. The presence of a context phoneme, however, attenuated EFRs by a degree similar to that caused by adaptation (~21-23 nV). Such a trade-off in the short- and long-term influence of context phonemes suggests that the benefit of interleaving EFR-eliciting vowels with other context phonemes depends on whether the use of consonant-vowel syllables is critical to improve the validity of EFR applications.
Collapse
Affiliation(s)
- Vijayalakshmi Easwar
- Department of Communication Sciences & Disorders, University of Wisconsin-Madison, Madison, USA.,Waisman Center, University of Wisconsin-Madison, Madison, USA
| | - Lauren Chung
- Department of Communication Sciences & Disorders, University of Wisconsin-Madison, Madison, USA.,Waisman Center, University of Wisconsin-Madison, Madison, USA
| |
Collapse
|
6
|
Hennessy S, Mack WJ, Habibi A. Speech-in-noise perception in musicians and non-musicians: A multi-level meta-analysis. Hear Res 2022; 416:108442. [PMID: 35078132 DOI: 10.1016/j.heares.2022.108442] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/17/2021] [Revised: 01/10/2022] [Accepted: 01/13/2022] [Indexed: 01/25/2023]
Abstract
Speech-in-noise perception, the ability to hear a relevant voice within a noisy background, is important for successful communication. Musicians have been reported to perform better than non-musicians on speech-in-noise tasks. This meta-analysis uses a multi-level design to assess the claim that musicians have superior speech-in-noise abilities compared to non-musicians. Across 31 studies and 62 effect sizes, the overall effect of musician status on speech-in-noise ability is significant, with a moderate effect size (g = 0.58), 95% CI [0.42, 0.74]. The overall effect of musician status was not moderated by within-study IQ equivalence, target stimulus, target contextual information, type of background noise, or age. We conclude that musicians show superior speech-in-noise abilities compared to non-musicians, not modified by age, IQ, or speech task parameters. These effects may reflect changes due to music training or predisposed auditory advantages that encourage musicianship.
Collapse
Affiliation(s)
- Sarah Hennessy
- Brain and Creativity Institute, University of Southern California, Los Angeles, CA, United States
| | - Wendy J Mack
- Department of Population and Public Health Sciences, University of Southern California, Los Angeles, CA, United States
| | - Assal Habibi
- Brain and Creativity Institute, University of Southern California, Los Angeles, CA, United States.
| |
Collapse
|
7
|
Gertsovski A, Ahissar M. Reduced Learning of Sound Categories in Dyslexia Is Associated with Reduced Regularity-Induced Auditory Cortex Adaptation. J Neurosci 2022; 42:1328-1342. [PMID: 34969869 PMCID: PMC8883854 DOI: 10.1523/jneurosci.1533-21.2021] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2021] [Revised: 12/09/2021] [Accepted: 12/11/2021] [Indexed: 11/21/2022] Open
Abstract
A main characteristic of dyslexia is poor use of sound categories. We now studied within-session learning of new sound categories in dyslexia, behaviorally and neurally, using fMRI. Human participants (males and females) with and without dyslexia were asked to discriminate which of two serially-presented tones had a higher pitch. The task was administered in two protocols, with and without a repeated reference frequency. The reference condition introduces regularity, and enhances frequency sensitivity in typically developing (TD) individuals. Enhanced sensitivity facilitates the formation of "high" and "low" pitch categories above and below this reference, respectively. We found that in TDs, learning was paralleled by a gradual decrease in activation of the primary auditory cortex (PAC), and reduced activation of the superior temporal gyrus (STG) and left posterior parietal cortex (PPC), which are important for using sensory history. No such sensitivity was found among individuals with dyslexia (IDDs). Rather, IDDs showed reduced behavioral learning of stimulus regularities and no regularity-associated adaptation in the auditory cortex or in higher-level regions. We propose that IDDs' reduced cortical adaptation, associated with reduced behavioral learning of sound regularities, underlies their impoverished use of stimulus history, and consequently impedes their formation of rich sound categories.SIGNIFICANCE STATEMENT Reading difficulties in dyslexia are often attributed to poor use of phonological categories. To test whether poor category use could result from poor learning of new sound categories in general, we administered an auditory discrimination task that examined the learning of new pitch categories above and below a repeated reference sound. Individuals with dyslexia (IDDs) learned categories slower than typically developing (TD) individuals. TD individuals showed adaptation to the repeated sounds that paralleled the category learning in their primary auditory cortex (PAC) and other higher-level regions. In dyslexia, no brain region showed such adaptation. We suggest that poor learning of sound statistics in sensory regions may underlie the poor representations of both speech and nonspeech categories in dyslexia.
Collapse
Affiliation(s)
- Ayelet Gertsovski
- The Edmond and Lily Safra Center for Brain Sciences, The Hebrew University of Jerusalem, Jerusalem 9190401, Israel
| | - Merav Ahissar
- The Edmond and Lily Safra Center for Brain Sciences, The Hebrew University of Jerusalem, Jerusalem 9190401, Israel
- Department of Psychology, The Hebrew University of Jerusalem, Jerusalem 9190501, Israel
| |
Collapse
|
8
|
Homma NY, Bajo VM. Lemniscal Corticothalamic Feedback in Auditory Scene Analysis. Front Neurosci 2021; 15:723893. [PMID: 34489635 PMCID: PMC8417129 DOI: 10.3389/fnins.2021.723893] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2021] [Accepted: 07/30/2021] [Indexed: 12/15/2022] Open
Abstract
Sound information is transmitted from the ear to central auditory stations of the brain via several nuclei. In addition to these ascending pathways there exist descending projections that can influence the information processing at each of these nuclei. A major descending pathway in the auditory system is the feedback projection from layer VI of the primary auditory cortex (A1) to the ventral division of medial geniculate body (MGBv) in the thalamus. The corticothalamic axons have small glutamatergic terminals that can modulate thalamic processing and thalamocortical information transmission. Corticothalamic neurons also provide input to GABAergic neurons of the thalamic reticular nucleus (TRN) that receives collaterals from the ascending thalamic axons. The balance of corticothalamic and TRN inputs has been shown to refine frequency tuning, firing patterns, and gating of MGBv neurons. Therefore, the thalamus is not merely a relay stage in the chain of auditory nuclei but does participate in complex aspects of sound processing that include top-down modulations. In this review, we aim (i) to examine how lemniscal corticothalamic feedback modulates responses in MGBv neurons, and (ii) to explore how the feedback contributes to auditory scene analysis, particularly on frequency and harmonic perception. Finally, we will discuss potential implications of the role of corticothalamic feedback in music and speech perception, where precise spectral and temporal processing is essential.
Collapse
Affiliation(s)
- Natsumi Y. Homma
- Center for Integrative Neuroscience, University of California, San Francisco, San Francisco, CA, United States
- Coleman Memorial Laboratory, Department of Otolaryngology – Head and Neck Surgery, University of California, San Francisco, San Francisco, CA, United States
| | - Victoria M. Bajo
- Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
9
|
Zhang X, Gong Q. Context-dependent Plasticity and Strength of Subcortical Encoding of Musical Sounds Independently Underlie Pitch Discrimination for Music Melodies. Neuroscience 2021; 472:68-89. [PMID: 34358631 DOI: 10.1016/j.neuroscience.2021.07.032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2021] [Revised: 07/26/2021] [Accepted: 07/27/2021] [Indexed: 10/20/2022]
Abstract
Subcortical auditory nuclei contribute to pitch perception, but how subcortical sound encoding is related to pitch processing for music perception remains unclear. Conventionally, enhanced subcortical sound encoding is considered underlying superior pitch discrimination. However, associations between superior auditory perception and the context-dependent plasticity of subcortical sound encoding are also documented. Here, we explored the subcortical neural correlates to music pitch perception by analyzing frequency-following responses (FFRs) to musical sounds presented in a predictable context and a random context. We found that the FFR inter-trial phase-locking (ITPL) was negatively correlated with behavioral performances of discrimination of pitches in music melodies. It was also negatively correlated with the plasticity indices measuring the variability of FFRs to physically identical sounds between the two contexts. The plasticity indices were consistently positively correlated with pitch discrimination performances, suggesting the subcortical context-dependent plasticity underlying music pitch perception. Moreover, the raw FFR spectral strength was not significantly correlated with pitch discrimination performances. However, it was positively correlated with behavioral performances when the FFR ITPL was controlled by partial correlations, suggesting that the strength of subcortical sound encoding underlies music pitch perception. When the spectral strength was controlled by partial correlations, the negative ITPL-behavioral correlations were maintained. Furthermore, the FFR ITPL, the plasticity indices, and the FFR spectral strength were more correlated with pitch than with rhythm discrimination performances. These findings suggest that the context-dependent plasticity and the strength of subcortical encoding of musical sounds are independently and perhaps specifically associated with pitch perception for music melodies.
Collapse
Affiliation(s)
- Xiaochen Zhang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China; Shanghai Mental Health Center, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Qin Gong
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China; School of Medicine, Shanghai University, Shanghai, China.
| |
Collapse
|
10
|
Hennessy S, Wood A, Wilcox R, Habibi A. Neurophysiological improvements in speech-in-noise task after short-term choir training in older adults. Aging (Albany NY) 2021; 13:9468-9495. [PMID: 33824226 PMCID: PMC8064162 DOI: 10.18632/aging.202931] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Accepted: 03/26/2021] [Indexed: 01/24/2023]
Abstract
Perceiving speech in noise (SIN) is important for health and well-being and decreases with age. Musicians show improved speech-in-noise abilities and reduced age-related auditory decline, yet it is unclear whether short term music engagement has similar effects. In this randomized control trial we used a pre-post design to investigate whether a 12-week music intervention in adults aged 50-65 without prior music training and with subjective hearing loss improves well-being, speech-in-noise abilities, and auditory encoding and voluntary attention as indexed by auditory evoked potentials (AEPs) in a syllable-in-noise task, and later AEPs in an oddball task. Age and gender-matched adults were randomized to a choir or control group. Choir participants sang in a 2-hr ensemble with 1-hr home vocal training weekly; controls listened to a 3-hr playlist weekly, attended concerts, and socialized online with fellow participants. From pre- to post-intervention, no differences between groups were observed on quantitative measures of well-being or behavioral speech-in-noise abilities. In the choir group, but not the control group, changes in the N1 component were observed for the syllable-in-noise task, with increased N1 amplitude in the passive condition and decreased N1 latency in the active condition. During the oddball task, larger N1 amplitudes to the frequent standard stimuli were also observed in the choir but not control group from pre to post intervention. Findings have implications for the potential role of music training to improve sound encoding in individuals who are in the vulnerable age range and at risk of auditory decline.
Collapse
Affiliation(s)
- Sarah Hennessy
- Brain and Creativity Institute, University of Southern California, Los Angeles, CA 90089, USA
| | - Alison Wood
- Brain and Creativity Institute, University of Southern California, Los Angeles, CA 90089, USA
| | - Rand Wilcox
- Department of Psychology, University of Southern California, Los Angeles, CA 90089, USA
| | - Assal Habibi
- Brain and Creativity Institute, University of Southern California, Los Angeles, CA 90089, USA
| |
Collapse
|
11
|
Neural generators of the frequency-following response elicited to stimuli of low and high frequency: A magnetoencephalographic (MEG) study. Neuroimage 2021; 231:117866. [PMID: 33592244 DOI: 10.1016/j.neuroimage.2021.117866] [Citation(s) in RCA: 37] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2020] [Revised: 02/08/2021] [Accepted: 02/09/2021] [Indexed: 01/03/2023] Open
Abstract
The frequency-following response (FFR) to periodic complex sounds has gained recent interest in auditory cognitive neuroscience as it captures with great fidelity the tracking accuracy of the periodic sound features in the ascending auditory system. Seminal studies suggested the FFR as a correlate of subcortical sound encoding, yet recent studies aiming to locate its sources challenged this assumption, demonstrating that FFR receives some contribution from the auditory cortex. Based on frequency-specific phase-locking capabilities along the auditory hierarchy, we hypothesized that FFRs to higher frequencies would receive less cortical contribution than those to lower frequencies, hence supporting a major subcortical involvement for these high frequency sounds. Here, we used a magnetoencephalographic (MEG) approach to trace the neural sources of the FFR elicited in healthy adults (N = 19) to low (89 Hz) and high (333 Hz) frequency sounds. FFRs elicited to the high and low frequency sounds were clearly observable on MEG and comparable to those obtained in simultaneous electroencephalographic recordings. Distributed source modeling analyses revealed midbrain, thalamic, and cortical contributions to FFR, arranged in frequency-specific configurations. Our results showed that the main contribution to the high-frequency sound FFR originated in the inferior colliculus and the medial geniculate body of the thalamus, with no significant cortical contribution. In contrast, the low-frequency sound FFR had a major contribution located in the auditory cortices, and also received contributions originating in the midbrain and thalamic structures. These findings support the multiple generator hypothesis of the FFR and are relevant for our understanding of the neural encoding of sounds along the auditory hierarchy, suggesting a hierarchical organization of periodicity encoding.
Collapse
|
12
|
Meter enhances the subcortical processing of speech sounds at a strong beat. Sci Rep 2020; 10:15973. [PMID: 32994430 PMCID: PMC7525485 DOI: 10.1038/s41598-020-72714-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2019] [Accepted: 09/07/2020] [Indexed: 11/08/2022] Open
Abstract
The temporal structure of sound such as in music and speech increases the efficiency of auditory processing by providing listeners with a predictable context. Musical meter is a good example of a sound structure that is temporally organized in a hierarchical manner, with recent studies showing that meter optimizes neural processing, particularly for sounds located at a higher metrical position or strong beat. Whereas enhanced cortical auditory processing at times of high metric strength has been studied, there is to date no direct evidence showing metrical modulation of subcortical processing. In this work, we examined the effect of meter on the subcortical encoding of sounds by measuring human auditory frequency-following responses to speech presented at four different metrical positions. Results show that neural encoding of the fundamental frequency of the vowel was enhanced at the strong beat, and also that the neural consistency of the vowel was the highest at the strong beat. When comparing musicians to non-musicians, musicians were found, at the strong beat, to selectively enhance the behaviorally relevant component of the speech sound, namely the formant frequency of the transient part. Our findings indicate that the meter of sound influences subcortical processing, and this metrical modulation differs depending on musical expertise.
Collapse
|
13
|
Couth S, Prendergast G, Guest H, Munro KJ, Moore DR, Plack CJ, Ginsborg J, Dawes P. Investigating the effects of noise exposure on self-report, behavioral and electrophysiological indices of hearing damage in musicians with normal audiometric thresholds. Hear Res 2020; 395:108021. [PMID: 32631495 DOI: 10.1016/j.heares.2020.108021] [Citation(s) in RCA: 31] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/11/2020] [Revised: 05/02/2020] [Accepted: 06/11/2020] [Indexed: 01/11/2023]
Abstract
Musicians are at risk of hearing loss due to prolonged noise exposure, but they may also be at risk of early sub-clinical hearing damage, such as cochlear synaptopathy. In the current study, we investigated the effects of noise exposure on electrophysiological, behavioral and self-report correlates of hearing damage in young adult (age range = 18-27 years) musicians and non-musicians with normal audiometric thresholds. Early-career musicians (n = 76) and non-musicians (n = 47) completed a test battery including the Noise Exposure Structured Interview, pure-tone audiometry (PTA; 0.25-8 kHz), extended high-frequency (EHF; 12 and 16 kHz) thresholds, otoacoustic emissions (OAEs), auditory brainstem responses (ABRs), speech perception in noise (SPiN), and self-reported tinnitus, hyperacusis and hearing in noise difficulties. Total lifetime noise exposure was similar between musicians and non-musicians, the majority of which could be accounted for by recreational activities. Musicians showed significantly greater ABR wave I/V ratios than non-musicians and were also more likely to report experience of - and/or more severe - tinnitus, hyperacusis and hearing in noise difficulties, irrespective of noise exposure. A secondary analysis revealed that individuals with the highest levels of noise exposure had reduced outer hair cell function compared to individuals with the lowest levels of noise exposure, as measured by OAEs. OAE level was also related to PTA and EHF thresholds. High levels of noise exposure were also associated with a significant increase in ABR wave V latency, but only for males, and a higher prevalence and severity of hyperacusis. These findings suggest that there may be sub-clinical effects of noise exposure on various hearing metrics even at a relatively young age, but do not support a link between lifetime noise exposure and proxy measures of cochlear synaptopathy such as ABR wave amplitudes and SPiN. Closely monitoring OAEs, PTA and EHF thresholds when conventional PTA is within the clinically 'normal' range could provide a useful early metric of noise-induced hearing damage. This may be particularly relevant to early-career musicians as they progress through a period of intensive musical training, and thus interventions to protect hearing longevity may be vital.
Collapse
Affiliation(s)
- Samuel Couth
- Manchester Centre for Audiology and Deafness, University of Manchester, UK.
| | | | - Hannah Guest
- Manchester Centre for Audiology and Deafness, University of Manchester, UK
| | - Kevin J Munro
- Manchester Centre for Audiology and Deafness, University of Manchester, UK; Manchester Academic Health Science Centre, Manchester University Hospitals NHS Foundation Trust, UK
| | - David R Moore
- Manchester Centre for Audiology and Deafness, University of Manchester, UK; Communication Sciences Research Center, Cincinnati Children's Hospital Medical Centre, OH, USA
| | - Christopher J Plack
- Manchester Centre for Audiology and Deafness, University of Manchester, UK; Department of Psychology, Lancaster University, UK
| | | | - Piers Dawes
- Manchester Centre for Audiology and Deafness, University of Manchester, UK; Department of Linguistics, Macquarie University, Sydney, Australia
| |
Collapse
|
14
|
Font-Alaminos M, Ribas-Prats T, Gorina-Careta N, Escera C. Emergence of prediction error along the human auditory hierarchy. Hear Res 2020; 399:107954. [PMID: 32234254 DOI: 10.1016/j.heares.2020.107954] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/04/2019] [Revised: 03/13/2020] [Accepted: 03/17/2020] [Indexed: 11/25/2022]
Abstract
Auditory prediction errors have been extensively associated with the mismatch negativity (MMN), a cortical auditory evoked potential that denotes deviance detection. Yet, many studies lacked the appropriate controls to disentangle sensory adaptation from prediction error. Furthermore, subcortical deviance detection has been shown in humans through recordings of the frequency-following response (FFR), an early auditory evoked potential that reflects the neural tracking of the periodic characteristics of a sound, suggesting the possibility that prediction errors emerge subcortically in the auditory pathway. The present study aimed at investigating the emergence of prediction error along the auditory hierarchy in humans through combined recordings of the FFR and the MMN, tapping at subcortical and cortical levels, respectively, while disentangling prediction error from sensory adaptation with the use of appropriate controls. "Oddball" sequences of pure tones featuring repeated "standard" stimuli (269 Hz; p = 0.8) and rare "deviant" stimuli (p = 0.2; of 289, 329 and 409 Hz delivered in separated blocks to test "frequency separation" effects) were presented to nineteen healthy young participants. "Reversed" oddball sequences (where standard and deviant tones swapped their roles) were presented allowing comparison of responses to same physical stimuli as a function of functional role (i.e., standard, deviant). Critically, control sequences featuring five equiprobable tones (p = 0.2) allowed to dissociate sensory adaptation from prediction error. Results revealed that the MMN amplitude increased as a function of frequency separation yet displayed the same amplitude when retrieved against the control sequences, confirming previous results. FFRs showed repetition enhancement effects across all frequency separations, as supported by larger spectral amplitude to standard than to deviant and control stimuli. This pattern of results provides insights into the hierarchy of the human prediction error system in audition, suggesting that prediction errors in humans emerge at cortical level.
Collapse
Affiliation(s)
- Marta Font-Alaminos
- Brainlab-Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institute of Neurosciences, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institut de Recerca Sant Joan de Déu (IRSJD), Esplugues de Llobregat, Barcelona, Catalonia, Spain
| | - Teresa Ribas-Prats
- Brainlab-Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institute of Neurosciences, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institut de Recerca Sant Joan de Déu (IRSJD), Esplugues de Llobregat, Barcelona, Catalonia, Spain
| | - Natàlia Gorina-Careta
- Brainlab-Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institute of Neurosciences, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institut de Recerca Sant Joan de Déu (IRSJD), Esplugues de Llobregat, Barcelona, Catalonia, Spain
| | - Carles Escera
- Brainlab-Cognitive Neuroscience Research Group, Department of Clinical Psychology and Psychobiology, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institute of Neurosciences, University of Barcelona, Passeig de la Vall d'Hebron 171, 08035, Barcelona, Catalonia, Spain; Institut de Recerca Sant Joan de Déu (IRSJD), Esplugues de Llobregat, Barcelona, Catalonia, Spain.
| |
Collapse
|
15
|
Font-Alaminos M, Cornella M, Costa-Faidella J, Hervás A, Leung S, Rueda I, Escera C. Increased subcortical neural responses to repeating auditory stimulation in children with autism spectrum disorder. Biol Psychol 2020; 149:107807. [DOI: 10.1016/j.biopsycho.2019.107807] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2019] [Revised: 10/29/2019] [Accepted: 10/30/2019] [Indexed: 01/12/2023]
|
16
|
Same or different pitch? Effects of musical expertise, pitch difference, and auditory task on the pitch discrimination ability of musicians and non-musicians. Exp Brain Res 2019; 238:247-258. [PMID: 31844911 DOI: 10.1007/s00221-019-05707-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2019] [Accepted: 12/07/2019] [Indexed: 10/25/2022]
Abstract
Musical expertise promotes both the perception and the processing of music. The aim of the present study was to analyze if musicians compared to non-musicians already have auditory processing advantages at the neural level. 50 musicians and 50 non-musicians worked on a task to determine the individual auditory difference threshold (individual JND threshold). A passive oddball paradigm followed while the EEG activity was recorded. Frequent standard sounds (528 hertz [Hz]) and rare deviant sounds (individual JND threshold, 535 Hz, and 558 Hz) were presented in the oddball paradigm. The mismatch negativity (MMN) and the P3a were used as indicators of auditory discrimination skills for frequency differences. Musicians had significantly smaller individual JND thresholds than non-musicians, but musicians were not faster than non-musicians. Musicians and non-musicians showed both the MMN and the P3a at the 535 Hz and 558 Hz condition. In the individual JND threshold condition, non-musicians, whose individual JND threshold was at 539.8 Hz (and therefore even above the deviant sound of 535 Hz), predictably showed the MMN and the P3a. Musicians, whose individual JND threshold was at 531.1 Hz (and thus close to the standard sound of 528 Hz), showed no MMN and P3a-although they were behaviorally able to differentiate frequencies individually within their JND threshold range. This may indicate a key role of attention in triggering the MMN during the detection of frequency differences in the individual JND threshold range (see Tervaniemi et al. in Exp Brain 161:1-10, 2005).
Collapse
|
17
|
Rosenthal MA. A systematic review of the voice-tagging hypothesis of speech-in-noise perception. Neuropsychologia 2019; 136:107256. [PMID: 31715197 DOI: 10.1016/j.neuropsychologia.2019.107256] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2019] [Revised: 11/03/2019] [Accepted: 11/06/2019] [Indexed: 01/05/2023]
Abstract
The voice-tagging hypothesis claims that individuals who better represent pitch information in a speaker's voice, as measured with the frequency following response (FFR), will be better at speech-in-noise perception. The hypothesis has been provided to explain how music training might improve speech-in-noise perception. This paper reviews studies that are relevant to the voice-tagging hypothesis, including studies on musicians and nonmusicians. Most studies on musicians show greater f0 amplitude compared to controls. Most studies on nonmusicians do not show group differences in f0 amplitude. Across all studies reviewed, f0 amplitude does not consistently predict accuracy in speech-in-noise perception. The evidence suggests that music training does not improve speech-in-noise perception via enhanced subcortical representation of the f0.
Collapse
Affiliation(s)
- Matthew A Rosenthal
- University of Kansas, 1450 Jayhawk Blvd, Lawrence, KS, 66045, Department of Psychology, United States.
| |
Collapse
|
18
|
The effects of short-term musical training on the neural processing of speech-in-noise in older adults. Brain Cogn 2019; 136:103592. [PMID: 31404817 DOI: 10.1016/j.bandc.2019.103592] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2018] [Revised: 07/29/2019] [Accepted: 08/01/2019] [Indexed: 01/31/2023]
Abstract
Experienced musicians outperform non-musicians in understanding speech-in-noise (SPIN). The benefits of lifelong musicianship endure into older age, where musicians experience smaller declines in their ability to understand speech in noisy environments. However, it is presently unknown whether commencing musical training in old age can also counteract age-related decline in speech perception, and whether such training induces changes in neural processing of speech. Here, we recruited older adult non-musicians and assigned them to receive a short course of piano or videogame training, or no training. Participants completed two sessions of functional Magnetic Resonance Imaging where they performed a SPIN task prior to and following training. While we found no direct benefit of musical training upon SPIN perception, an exploratory Region of Interest analysis revealed increased cortical responses to speech in left Middle Frontal and Supramarginal Gyri which correlated with changes in SPIN task performance in the group which received music training. These results suggest that short-term musical training in older adults may enhance neural encoding of speech, with the potential to reduce age-related decline in speech perception.
Collapse
|
19
|
BinKhamis G, Perugia E, O'Driscoll M, Kluk K. Speech-ABRs in cochlear implant recipients: feasibility study. Int J Audiol 2019; 58:678-684. [PMID: 31132012 DOI: 10.1080/14992027.2019.1619100] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
Objective: The aim of this study was to assess the feasibility of recording speech-ABRs from cochlear implant (CI) recipients, and to remove the artefact using a clinically applicable single-channel approach. Design: Speech-ABRs were recorded to a 40 ms [da] presented via loudspeaker using a two-channel electrode montage. Additionally, artefacts were recorded using an artificial-head incorporating a MED-EL CI with stimulation parameters as similar as possible to those of three MED-EL participants. A single-channel artefact removal technique was applied to all responses. Study sample: A total of 12 adult CI recipients (6 Cochlear Nucleus and 6 MED-EL CIs). Results: Responses differed according to the CI type, artefact removal resulted in responses containing speech-ARB characteristics in two MED-EL CI participants; however, it was not possible to verify whether these were true responses or were modulated by artefacts, and artefact removal was successful from the artificial-head recordings. Conclusions: This is the first study that attempted to record speech-ABRs from CI recipients. Results suggest that there is a potential for application of a single-channel approach to artefact removal. However, a more robust and adaptive approach to artefact removal that includes a method to verify true responses is needed.
Collapse
Affiliation(s)
- Ghada BinKhamis
- Manchester Centre for Audiology and Deafness, Manchester Academic Health Science Centre, University of Manchester , Manchester , UK.,King Fahad Medical City , Riyadh , Saudi Arabia
| | - Emanuele Perugia
- Manchester Centre for Audiology and Deafness, Manchester Academic Health Science Centre, University of Manchester , Manchester , UK
| | - Martin O'Driscoll
- Manchester Centre for Audiology and Deafness, Manchester Academic Health Science Centre, University of Manchester , Manchester , UK.,Manchester Auditory Implant Centre, Manchester University Hospitals NHS Foundation Trust , Manchester , UK
| | - Karolina Kluk
- Manchester Centre for Audiology and Deafness, Manchester Academic Health Science Centre, University of Manchester , Manchester , UK
| |
Collapse
|
20
|
Morse-Fortier C, Parrish MM, Baran JA, Freyman RL. The Effects of Musical Training on Speech Detection in the Presence of Informational and Energetic Masking. Trends Hear 2019; 21:2331216517739427. [PMID: 29161982 PMCID: PMC5703091 DOI: 10.1177/2331216517739427] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022] Open
Abstract
Recent research has suggested that musicians have an advantage in some speech-in-noise paradigms, but not all. Whether musicians outperform nonmusicians on a given speech-in-noise task may well depend on the type of noise involved. To date, few groups have specifically studied the role that informational masking plays in the observation of a musician advantage. The current study investigated the effect of musicianship on listeners’ ability to overcome informational versus energetic masking of speech. Monosyllabic words were presented in four conditions that created similar energetic masking but either high or low informational masking. Two of these conditions used noise-vocoded target and masking stimuli to determine whether the absence of natural fine structure and spectral variations influenced any musician advantage. Forty young normal-hearing listeners (20 musicians and 20 nonmusicians) completed the study. There was a significant overall effect of participant group collapsing across the four conditions; however, planned comparisons showed musicians’ thresholds were only significantly better in the high informational masking natural speech condition, where the musician advantage was approximately 3 dB. These results add to the mounting evidence that informational masking plays a role in the presence and amount of musician benefit.
Collapse
Affiliation(s)
| | - Mary M Parrish
- 1 Department of Communication Disorders, University of Massachusetts Amherst, MA, USA
| | - Jane A Baran
- 1 Department of Communication Disorders, University of Massachusetts Amherst, MA, USA
| | - Richard L Freyman
- 1 Department of Communication Disorders, University of Massachusetts Amherst, MA, USA
| |
Collapse
|
21
|
Interactive effects of linguistic abstraction and stimulus statistics in the online modulation of neural speech encoding. Atten Percept Psychophys 2018; 81:1020-1033. [PMID: 30565097 DOI: 10.3758/s13414-018-1621-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Speech processing is highly modulated by context. Prior studies examining frequency-following responses (FFRs), an electrophysiological 'neurophonic' potential that faithfully reflects phase-locked activity from neural ensembles within the auditory network, have demonstrated that stimulus context modulates the integrity of speech encoding. The extent to which context-dependent encoding reflects general auditory properties or interactivities between statistical and higher-level linguistic processes remains unexplored. Our study examined whether speech encoding, as reflected by FFRs, is modulated by abstract phonological relationships between a stimulus and surrounding contexts. FFRs were elicited to a Mandarin rising-tone syllable (/ji-TR/, 'second') randomly presented with other syllables in three contexts from 17 native listeners. In a contrastive context, /ji-TR/ occurred with meaning-contrastive high-level-tone syllables (/ji-H/, 'one'). In an allotone context, TR occurred with dipping-tone syllables /ji-D/, a non-meaning-contrastive variant of /ji-TR/. In a repetitive context, the same /ji-TR/ occurred with other speech tokens of /ji-TR/. Consistent with prior work, neural tracking of /ji-TR/ pitch contour was more faithful in the repetitive condition wherein /ji-TR/ occurred more predictably (p = 1) than in the contrastive condition (p = 0.34). Crucially, in the allotone context, neural tracking of /ji-TR/ was more accurate relative to the contrastive context, despite both having an identical transitional probability (p = 0.34). Mechanistically, the non-meaning-contrastive relationship may have augmented the probability to /ji-TR/ occurrence in the allotone context. Results indicate online interactions between bottom-up and top-down mechanisms, which facilitate speech perception. Such interactivities may predictively fine-tune incoming speech encoding using linguistic and statistical information from prior context.
Collapse
|
22
|
Dhatri SD, Gnanateja GN, Kumar UA, Maruthy S. Gender-bias in the sensory representation of infant cry. Neurosci Lett 2018; 678:138-143. [DOI: 10.1016/j.neulet.2018.04.043] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2017] [Revised: 04/09/2018] [Accepted: 04/23/2018] [Indexed: 10/17/2022]
|
23
|
Elmer S, Jäncke L. Relationships between music training, speech processing, and word learning: a network perspective. Ann N Y Acad Sci 2018; 1423:10-18. [PMID: 29542125 DOI: 10.1111/nyas.13581] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2017] [Revised: 11/16/2017] [Accepted: 11/27/2017] [Indexed: 01/19/2023]
Abstract
Numerous studies have documented the behavioral advantages conferred on professional musicians and children undergoing music training in processing speech sounds varying in the spectral and temporal dimensions. These beneficial effects have previously often been associated with local functional and structural changes in the auditory cortex (AC). However, this perspective is oversimplified, in that it does not take into account the intrinsic organization of the human brain, namely, neural networks and oscillatory dynamics. Therefore, we propose a new framework for extending these previous findings to a network perspective by integrating multimodal imaging, electrophysiology, and neural oscillations. In particular, we provide concrete examples of how functional and structural connectivity can be used to model simple neural circuits exerting a modulatory influence on AC activity. In addition, we describe how such a network approach can be used for better comprehending the beneficial effects of music training on more complex speech functions, such as word learning.
Collapse
Affiliation(s)
- Stefan Elmer
- Division of Neuropsychology (Auditory Research Group Zurich, ARGZ), Institute of Psychology, University of Zurich, Zurich, Switzerland
| | - Lutz Jäncke
- Division of Neuropsychology (Auditory Research Group Zurich, ARGZ), Institute of Psychology, University of Zurich, Zurich, Switzerland
- Center for Integrative Human Physiology (ZIHP), University of Zurich, Zurich, Switzerland
- International Normal Aging and Plasticity Imaging Center (INAPIC), University of Zurich, Zurich, Switzerland
- University Research Priority Program (URPP) "Dynamic of Healthy Aging", University of Zurich, Zurich, Switzerland
- Department of Special Education, King Abdulaziz University, Jeddah, Saudi Arabia
| |
Collapse
|
24
|
Irvine DRF. Auditory perceptual learning and changes in the conceptualization of auditory cortex. Hear Res 2018; 366:3-16. [PMID: 29551308 DOI: 10.1016/j.heares.2018.03.011] [Citation(s) in RCA: 33] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/13/2017] [Revised: 03/06/2018] [Accepted: 03/09/2018] [Indexed: 12/11/2022]
Abstract
Perceptual learning, improvement in discriminative ability as a consequence of training, is one of the forms of sensory system plasticity that has driven profound changes in our conceptualization of sensory cortical function. Psychophysical and neurophysiological studies of auditory perceptual learning have indicated that the characteristics of the learning, and by implication the nature of the underlying neural changes, are highly task specific. Some studies in animals have indicated that recruitment of neurons to the population responding to the training stimuli, and hence an increase in the so-called cortical "area of representation" of those stimuli, is the substrate of improved performance, but such changes have not been observed in other studies. A possible reconciliation of these conflicting results is provided by evidence that changes in area of representation constitute a transient stage in the processes underlying perceptual learning. This expansion - renormalization hypothesis is supported by evidence from studies of the learning of motor skills, another form of procedural learning, but leaves open the nature of the permanent neural substrate of improved performance. Other studies have suggested that the substrate might be reduced response variability - a decrease in internal noise. Neuroimaging studies in humans have also provided compelling evidence that training results in long-term changes in auditory cortical function and in the auditory brainstem frequency-following response. Musical training provides a valuable model, but the evidence it provides is qualified by the fact that most such training is multimodal and sensorimotor, and that few of the studies are experimental and allow control over confounding variables. More generally, the overwhelming majority of experimental studies of the various forms of auditory perceptual learning have established the co-occurrence of neural and perceptual changes, but have not established that the former are causally related to the latter. Important forms of perceptual learning in humans are those involved in language acquisition and in the improvement in speech perception performance of post-lingually deaf cochlear implantees over the months following implantation. The development of a range of auditory training programs has focused interest on the factors determining the extent to which perceptual learning is specific or generalises to tasks other than those used in training. The context specificity demonstrated in a number of studies of perceptual learning suggests a multiplexing model, in which learning relating to a particular stimulus attribute depends on a subset of the diverse inputs to a given cortical neuron being strengthened, and different subsets being gated by top-down influences. This hypothesis avoids the difficulty of balancing system stability with plasticity, which is a problem for recruitment hypotheses. The characteristics of auditory perceptual learning reflect the fact that auditory cortex forms part of distributed networks that integrate the representation of auditory stimuli with attention, decision, and reward processes.
Collapse
Affiliation(s)
- Dexter R F Irvine
- Bionics Institute, East Melbourne, Victoria 3002, Australia; School of Psychological Sciences, Monash University, Victoria 3800, Australia.
| |
Collapse
|
25
|
Madsen SMK, Whiteford KL, Oxenham AJ. Musicians do not benefit from differences in fundamental frequency when listening to speech in competing speech backgrounds. Sci Rep 2017; 7:12624. [PMID: 28974705 PMCID: PMC5626707 DOI: 10.1038/s41598-017-12937-9] [Citation(s) in RCA: 43] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2017] [Accepted: 09/11/2017] [Indexed: 11/09/2022] Open
Abstract
Recent studies disagree on whether musicians have an advantage over non-musicians in understanding speech in noise. However, it has been suggested that musicians may be able to use differences in fundamental frequency (F0) to better understand target speech in the presence of interfering talkers. Here we studied a relatively large (N = 60) cohort of young adults, equally divided between non-musicians and highly trained musicians, to test whether the musicians were better able to understand speech either in noise or in a two-talker competing speech masker. The target speech and competing speech were presented with either their natural F0 contours or on a monotone F0, and the F0 difference between the target and masker was systematically varied. As expected, speech intelligibility improved with increasing F0 difference between the target and the two-talker masker for both natural and monotone speech. However, no significant intelligibility advantage was observed for musicians over non-musicians in any condition. Although F0 discrimination was significantly better for musicians than for non-musicians, it was not correlated with speech scores. Overall, the results do not support the hypothesis that musical training leads to improved speech intelligibility in complex speech or noise backgrounds.
Collapse
Affiliation(s)
- Sara M K Madsen
- Hearing Systems, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads 352, 2800, Kgs. Lyngby, Denmark.
| | - Kelly L Whiteford
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, MN, 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, 75 East River Parkway, Minneapolis, MN, 55455, USA
| |
Collapse
|
26
|
Coffey EBJ, Chepesiuk AMP, Herholz SC, Baillet S, Zatorre RJ. Neural Correlates of Early Sound Encoding and their Relationship to Speech-in-Noise Perception. Front Neurosci 2017; 11:479. [PMID: 28890684 PMCID: PMC5575455 DOI: 10.3389/fnins.2017.00479] [Citation(s) in RCA: 49] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2017] [Accepted: 08/11/2017] [Indexed: 01/05/2023] Open
Abstract
Speech-in-noise (SIN) perception is a complex cognitive skill that affects social, vocational, and educational activities. Poor SIN ability particularly affects young and elderly populations, yet varies considerably even among healthy young adults with normal hearing. Although SIN skills are known to be influenced by top-down processes that can selectively enhance lower-level sound representations, the complementary role of feed-forward mechanisms and their relationship to musical training is poorly understood. Using a paradigm that minimizes the main top-down factors that have been implicated in SIN performance such as working memory, we aimed to better understand how robust encoding of periodicity in the auditory system (as measured by the frequency-following response) contributes to SIN perception. Using magnetoencephalograpy, we found that the strength of encoding at the fundamental frequency in the brainstem, thalamus, and cortex is correlated with SIN accuracy. The amplitude of the slower cortical P2 wave was previously also shown to be related to SIN accuracy and FFR strength; we use MEG source localization to show that the P2 wave originates in a temporal region anterior to that of the cortical FFR. We also confirm that the observed enhancements were related to the extent and timing of musicianship. These results are consistent with the hypothesis that basic feed-forward sound encoding affects SIN perception by providing better information to later processing stages, and that modifying this process may be one mechanism through which musical training might enhance the auditory networks that subserve both musical and language functions.
Collapse
Affiliation(s)
- Emily B J Coffey
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada
| | - Alexander M P Chepesiuk
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada
| | - Sibylle C Herholz
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada.,German Center for Neurodegenerative DiseasesBonn, Germany
| | - Sylvain Baillet
- Centre for Research on Brain, Language and MusicMontréal, QC, Canada.,McConnell Brain Imaging Centre, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada
| | - Robert J Zatorre
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill UniversityMontréal, QC, Canada.,Laboratory for Brain, Music and Sound ResearchMontréal, QC, Canada.,Centre for Research on Brain, Language and MusicMontréal, QC, Canada
| |
Collapse
|
27
|
Perrachione TK, Del Tufo SN, Winter R, Murtagh J, Cyr A, Chang P, Halverson K, Ghosh SS, Christodoulou JA, Gabrieli JDE. Dysfunction of Rapid Neural Adaptation in Dyslexia. Neuron 2017; 92:1383-1397. [PMID: 28009278 DOI: 10.1016/j.neuron.2016.11.020] [Citation(s) in RCA: 85] [Impact Index Per Article: 12.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2016] [Revised: 08/16/2016] [Accepted: 10/26/2016] [Indexed: 11/16/2022]
Abstract
Identification of specific neurophysiological dysfunctions resulting in selective reading difficulty (dyslexia) has remained elusive. In addition to impaired reading development, individuals with dyslexia frequently exhibit behavioral deficits in perceptual adaptation. Here, we assessed neurophysiological adaptation to stimulus repetition in adults and children with dyslexia for a wide variety of stimuli, spoken words, written words, visual objects, and faces. For every stimulus type, individuals with dyslexia exhibited significantly diminished neural adaptation compared to controls in stimulus-specific cortical areas. Better reading skills in adults and children with dyslexia were associated with greater repetition-induced neural adaptation. These results highlight a dysfunction of rapid neural adaptation as a core neurophysiological difference in dyslexia that may underlie impaired reading development. Reduced neurophysiological adaptation may relate to prior reports of reduced behavioral adaptation in dyslexia and may reveal a difference in brain functions that ultimately results in a specific reading impairment.
Collapse
Affiliation(s)
- Tyler K Perrachione
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.
| | - Stephanie N Del Tufo
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Rebecca Winter
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Jack Murtagh
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Abigail Cyr
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Patricia Chang
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Kelly Halverson
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Satrajit S Ghosh
- Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Joanna A Christodoulou
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - John D E Gabrieli
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.
| |
Collapse
|
28
|
Heald SLM, Van Hedger SC, Nusbaum HC. Perceptual Plasticity for Auditory Object Recognition. Front Psychol 2017; 8:781. [PMID: 28588524 PMCID: PMC5440584 DOI: 10.3389/fpsyg.2017.00781] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2016] [Accepted: 04/26/2017] [Indexed: 01/25/2023] Open
Abstract
In our auditory environment, we rarely experience the exact acoustic waveform twice. This is especially true for communicative signals that have meaning for listeners. In speech and music, the acoustic signal changes as a function of the talker (or instrument), speaking (or playing) rate, and room acoustics, to name a few factors. Yet, despite this acoustic variability, we are able to recognize a sentence or melody as the same across various kinds of acoustic inputs and determine meaning based on listening goals, expectations, context, and experience. The recognition process relates acoustic signals to prior experience despite variability in signal-relevant and signal-irrelevant acoustic properties, some of which could be considered as "noise" in service of a recognition goal. However, some acoustic variability, if systematic, is lawful and can be exploited by listeners to aid in recognition. Perceivable changes in systematic variability can herald a need for listeners to reorganize perception and reorient their attention to more immediately signal-relevant cues. This view is not incorporated currently in many extant theories of auditory perception, which traditionally reduce psychological or neural representations of perceptual objects and the processes that act on them to static entities. While this reduction is likely done for the sake of empirical tractability, such a reduction may seriously distort the perceptual process to be modeled. We argue that perceptual representations, as well as the processes underlying perception, are dynamically determined by an interaction between the uncertainty of the auditory signal and constraints of context. This suggests that the process of auditory recognition is highly context-dependent in that the identity of a given auditory object may be intrinsically tied to its preceding context. To argue for the flexible neural and psychological updating of sound-to-meaning mappings across speech and music, we draw upon examples of perceptual categories that are thought to be highly stable. This framework suggests that the process of auditory recognition cannot be divorced from the short-term context in which an auditory object is presented. Implications for auditory category acquisition and extant models of auditory perception, both cognitive and neural, are discussed.
Collapse
|
29
|
Maruthy S, Kumar UA, Gnanateja GN. Functional Interplay Between the Putative Measures of Rostral and Caudal Efferent Regulation of Speech Perception in Noise. J Assoc Res Otolaryngol 2017; 18:635-648. [PMID: 28447225 DOI: 10.1007/s10162-017-0623-y] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2016] [Accepted: 03/22/2017] [Indexed: 01/23/2023] Open
Abstract
Efferent modulation has been demonstrated to be very important for speech perception, especially in the presence of noise. We examined the functional relationship between two efferent systems: the rostral and caudal efferent pathways and their individual influences on speech perception in noise. Earlier studies have shown that these two efferent mechanisms were correlated with speech perception in noise. However, previously, these mechanisms were studied in isolation, and their functional relationship with each other was not investigated. We used a correlational design to study the relationship if any, between these two mechanisms in young and old normal hearing individuals. We recorded context-dependent brainstem encoding as an index of rostral efferent function and contralateral suppression of otoacoustic emissions as an index of caudal efferent function in groups with good and poor speech perception in noise. These efferent mechanisms were analysed for their relationship with each other and with speech perception in noise. We found that the two efferent mechanisms did not show any functional relationship. Interestingly, both the efferent mechanisms correlated with speech perception in noise and they even emerged as significant predictors. Based on the data, we posit that the two efferent mechanisms function relatively independently but with a common goal of fine-tuning the afferent input and refining auditory perception in degraded listening conditions.
Collapse
Affiliation(s)
- Sandeep Maruthy
- Electrophysiology Laboratory, Department of Audiology, All India Institute of Speech and Hearing, Manasagangothri, Mysore, Karnataka, IN-570006, India
| | - U Ajith Kumar
- Electrophysiology Laboratory, Department of Audiology, All India Institute of Speech and Hearing, Manasagangothri, Mysore, Karnataka, IN-570006, India
| | - G Nike Gnanateja
- Electrophysiology Laboratory, Department of Audiology, All India Institute of Speech and Hearing, Manasagangothri, Mysore, Karnataka, IN-570006, India.
| |
Collapse
|
30
|
Speech-in-noise perception in musicians: A review. Hear Res 2017; 352:49-69. [PMID: 28213134 DOI: 10.1016/j.heares.2017.02.006] [Citation(s) in RCA: 83] [Impact Index Per Article: 11.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/26/2016] [Revised: 02/01/2017] [Accepted: 02/05/2017] [Indexed: 11/23/2022]
Abstract
The ability to understand speech in the presence of competing sound sources is an important neuroscience question in terms of how the nervous system solves this computational problem. It is also a critical clinical problem that disproportionally affects the elderly, children with language-related learning disorders, and those with hearing loss. Recent evidence that musicians have an advantage on this multifaceted skill has led to the suggestion that musical training might be used to improve or delay the decline of speech-in-noise (SIN) function. However, enhancements have not been universally reported, nor have the relative contributions of different bottom-up versus top-down processes, and their relation to preexisting factors been disentangled. This information that would be helpful to establish whether there is a real effect of experience, what exactly is its nature, and how future training-based interventions might target the most relevant components of cognitive processes. These questions are complicated by important differences in study design and uneven coverage of neuroimaging modality. In this review, we aim to systematize recent results from studies that have specifically looked at musician-related differences in SIN by their study design properties, to summarize the findings, and to identify knowledge gaps for future work.
Collapse
|
31
|
Jaffe-Dax S, Frenkel O, Ahissar M. Dyslexics' faster decay of implicit memory for sounds and words is manifested in their shorter neural adaptation. eLife 2017; 6. [PMID: 28115055 PMCID: PMC5279949 DOI: 10.7554/elife.20557] [Citation(s) in RCA: 31] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2016] [Accepted: 01/09/2017] [Indexed: 11/13/2022] Open
Abstract
Dyslexia is a prevalent reading disability whose underlying mechanisms are still disputed. We studied the neural mechanisms underlying dyslexia using a simple frequency-discrimination task. Though participants were asked to compare the two tones in each trial, implicit memory of previous trials affected their responses. We hypothesized that implicit memory decays faster among dyslexics. We tested this by increasing the temporal intervals between consecutive trials, and by measuring the behavioral impact and ERP responses from the auditory cortex. Dyslexics showed a faster decay of implicit memory effects on both measures, with similar time constants. Finally, faster decay of implicit memory also characterized the impact of sound regularities in benefitting dyslexics' oral reading rate. Their benefit decreased faster as a function of the time interval from the previous reading of the same non-word. We propose that dyslexics’ shorter neural adaptation paradoxically accounts for their longer reading times, since it reduces their temporal window of integration of past stimuli, resulting in noisier and less reliable predictions for both simple and complex stimuli. Less reliable predictions limit their acquisition of reading expertise. DOI:http://dx.doi.org/10.7554/eLife.20557.001 The term “dyslexia” comes from the Greek for “difficulty with words”. People with dyslexia struggle with reading and spelling: they may mix up letters within words and tend to read and write more slowly than others. However, not every symptom of dyslexia is related to literacy. Affected individuals also differ from good readers on simple perceptual tasks, such as distinguishing between tones of different frequencies. In a series of trials involving discrimination between pairs of tones, a person’s performance on each trial will be influenced by the tones presented on previous trials. Both good readers and individuals with dyslexia automatically form a subconscious memory of the tones they hear, and use this memory to guide their performance on subsequent trials. However, people with dyslexia benefit less from this effect than good readers. Jaffe-Dax et al. have now identified the mechanism that underlies this phenomenon, revealing new insights into how dyslexia influences brain activity. By varying the interval between successive pairs of tones, the experiments showed that the memory of previous tones decays faster in people with dyslexia than in good readers. A similar effect occurs when the stimuli are nonsense words. Both good and poor readers manage to read nonsense words more quickly on their second attempt. However, people with dyslexia benefit less from the previous exposure when the gap between repetitions is longer than a couple of seconds. Further studies are needed to determine whether and how the faster decay of memory traces for words is related to impaired reading ability in people with dyslexia. One possibility is that the faster decay of memory traces makes it more difficult to predict future stimuli, which may impair reading. An imaging study is underway to investigate where in the brain this rapid decay of memory traces occurs. DOI:http://dx.doi.org/10.7554/eLife.20557.002
Collapse
Affiliation(s)
- Sagi Jaffe-Dax
- Edmond and Lily Safra Center for Brain Sciences, The Hebrew University of Jerusalem, Jerusalem, Israel.,Department of Psychology, Princeton University, Princeton, United States
| | - Or Frenkel
- Psychology Department, The Hebrew University of Jerusalem, Jerusalem, Israel
| | - Merav Ahissar
- Edmond and Lily Safra Center for Brain Sciences, The Hebrew University of Jerusalem, Jerusalem, Israel.,Psychology Department, The Hebrew University of Jerusalem, Jerusalem, Israel
| |
Collapse
|
32
|
The Role of the Auditory Brainstem in Regularity Encoding and Deviance Detection. THE FREQUENCY-FOLLOWING RESPONSE 2017. [DOI: 10.1007/978-3-319-47944-6_5] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2022]
|
33
|
Perrachione TK, Del Tufo SN, Winter R, Murtagh J, Cyr A, Chang P, Halverson K, Ghosh SS, Christodoulou JA, Gabrieli JDE. Dysfunction of Rapid Neural Adaptation in Dyslexia. Neuron 2016. [PMID: 28009278 DOI: 10.1016/j.neuron.2016.11.020"] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/30/2022]
Abstract
Identification of specific neurophysiological dysfunctions resulting in selective reading difficulty (dyslexia) has remained elusive. In addition to impaired reading development, individuals with dyslexia frequently exhibit behavioral deficits in perceptual adaptation. Here, we assessed neurophysiological adaptation to stimulus repetition in adults and children with dyslexia for a wide variety of stimuli, spoken words, written words, visual objects, and faces. For every stimulus type, individuals with dyslexia exhibited significantly diminished neural adaptation compared to controls in stimulus-specific cortical areas. Better reading skills in adults and children with dyslexia were associated with greater repetition-induced neural adaptation. These results highlight a dysfunction of rapid neural adaptation as a core neurophysiological difference in dyslexia that may underlie impaired reading development. Reduced neurophysiological adaptation may relate to prior reports of reduced behavioral adaptation in dyslexia and may reveal a difference in brain functions that ultimately results in a specific reading impairment.
Collapse
Affiliation(s)
- Tyler K Perrachione
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.
| | - Stephanie N Del Tufo
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Rebecca Winter
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Jack Murtagh
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Abigail Cyr
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Patricia Chang
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Kelly Halverson
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Satrajit S Ghosh
- Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Joanna A Christodoulou
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - John D E Gabrieli
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.
| |
Collapse
|
34
|
Zaballos MTP, Plasencia DP, González MLZ, de Miguel AR, Macías ÁR. Air traffic controllers' long-term speech-in-noise training effects: A control group study. Noise Health 2016; 18:376-381. [PMID: 27991470 PMCID: PMC5227019 DOI: 10.4103/1463-1741.195804] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022] Open
Abstract
Introduction: Speech perception in noise relies on the capacity of the auditory system to process complex sounds using sensory and cognitive skills. The possibility that these can be trained during adulthood is of special interest in auditory disorders, where speech in noise perception becomes compromised. Air traffic controllers (ATC) are constantly exposed to radio communication, a situation that seems to produce auditory learning. The objective of this study has been to quantify this effect. Subjects and Methods: 19 ATC and 19 normal hearing individuals underwent a speech in noise test with three signal to noise ratios: 5, 0 and −5 dB. Noise and speech were presented through two different loudspeakers in azimuth position. Speech tokes were presented at 65 dB SPL, while white noise files were at 60, 65 and 70 dB respectively. Results: Air traffic controllers outperform the control group in all conditions [P<0.05 in ANOVA and Mann-Whitney U tests]. Group differences were largest in the most difficult condition, SNR=−5 dB. However, no correlation between experience and performance were found for any of the conditions tested. The reason might be that ceiling performance is achieved much faster than the minimum experience time recorded, 5 years, although intrinsic cognitive abilities cannot be disregarded. Discussion: ATC demonstrated enhanced ability to hear speech in challenging listening environments. This study provides evidence that long-term auditory training is indeed useful in achieving better speech-in-noise understanding even in adverse conditions, although good cognitive qualities are likely to be a basic requirement for this training to be effective. Conclusion: Our results show that ATC outperform the control group in all conditions. Thus, this study provides evidence that long-term auditory training is indeed useful in achieving better speech-in-noise understanding even in adverse conditions.
Collapse
Affiliation(s)
- Maria T P Zaballos
- Laboratorio de Psicoacústica, Complejo Hospitalario Universitario Insular Materno Infantil, Las Palmas de Gran Canaria, Las Palmas, Spain
| | - Daniel P Plasencia
- ENT Department & Departamento de CC Quirúrgicas, Universidad de Las Palmas de Gran Canaria, Complejo Hospitalario Universitario Insular Materno Infantil, Las Palmas de Gran Canaria, Las Palmas, Spain
| | - María L Z González
- ENT Department, Complejo Hospitalario Universitario Insular Materno Infantil, Las Palmas de Gran Canaria, Las Palmas, Spain
| | - Angel R de Miguel
- Instituto Universitario de Sistemas Inteligentes y Aplicaciones Numéricas, Universidad de Las Palmas de Gran Canaria, Las Palmas de Gran Canaria, Las Palmas, Spain
| | - Ángel R Macías
- ENT Department & Departamento de CC Quirúrgicas, Universidad de Las Palmas de Gran Canaria, Complejo Hospitalario Universitario Insular Materno Infantil, Las Palmas de Gran Canaria, Las Palmas, Spain
| |
Collapse
|
35
|
Musicians' edge: A comparison of auditory processing, cognitive abilities and statistical learning. Hear Res 2016; 342:112-123. [DOI: 10.1016/j.heares.2016.10.008] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/29/2015] [Revised: 10/11/2016] [Accepted: 10/15/2016] [Indexed: 11/19/2022]
|
36
|
Lau JCY, Wong PCM, Chandrasekaran B. Context-dependent plasticity in the subcortical encoding of linguistic pitch patterns. J Neurophysiol 2016; 117:594-603. [PMID: 27832606 DOI: 10.1152/jn.00656.2016] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2016] [Accepted: 11/07/2016] [Indexed: 01/08/2023] Open
Abstract
We examined the mechanics of online experience-dependent auditory plasticity by assessing the influence of prior context on the frequency-following responses (FFRs), which reflect phase-locked responses from neural ensembles within the subcortical auditory system. FFRs were elicited to a Cantonese falling lexical pitch pattern from 24 native speakers of Cantonese in a variable context, wherein the falling pitch pattern randomly occurred in the context of two other linguistic pitch patterns; in a patterned context, wherein, the falling pitch pattern was presented in a predictable sequence along with two other pitch patterns, and in a repetitive context, wherein the falling pitch pattern was presented with 100% probability. We found that neural tracking of the stimulus pitch contour was most faithful and accurate when listening context was patterned and least faithful when the listening context was variable. The patterned context elicited more robust pitch tracking relative to the repetitive context, suggesting that context-dependent plasticity is most robust when the context is predictable but not repetitive. Our study demonstrates a robust influence of prior listening context that works to enhance online neural encoding of linguistic pitch patterns. We interpret these results as indicative of an interplay between contextual processes that are responsive to predictability as well as novelty in the presentation context. NEW & NOTEWORTHY Human auditory perception in dynamic listening environments requires fine-tuning of sensory signal based on behaviorally relevant regularities in listening context, i.e., online experience-dependent plasticity. Our finding suggests what partly underlie online experience-dependent plasticity are interplaying contextual processes in the subcortical auditory system that are responsive to predictability as well as novelty in listening context. These findings add to the literature that looks to establish the neurophysiological bases of auditory system plasticity, a central issue in auditory neuroscience.
Collapse
Affiliation(s)
- Joseph C Y Lau
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Shatin, Hong Kong
| | - Patrick C M Wong
- Department of Linguistics and Modern Languages, The Chinese University of Hong Kong, Shatin, Hong Kong.,Brain and Mind Institute, The Chinese University of Hong Kong, Shatin, Hong Kong
| | - Bharath Chandrasekaran
- Department of Communication Sciences and Disorders, Moody College of Communication, The University of Texas at Austin, Austin, Texas; .,Department of Psychology, College of Liberal Arts, The University of Texas at Austin, Austin, Texas.,Department of Linguistics, College of Liberal Arts, The University of Texas at Austin, Austin, Texas.,Institute of Mental Health Research, College of Liberal Arts, The University of Texas at Austin, Austin, Texas; and.,Institute for Neuroscience, The University of Texas at Austin, Austin, Texas
| |
Collapse
|
37
|
Kumar P, Anil SP, Grover V, Sanju HK, Sinha S. Cortical and subcortical processing of short duration speech stimuli in trained rock musicians: a pilot study. Eur Arch Otorhinolaryngol 2016; 274:1153-1160. [PMID: 27565157 DOI: 10.1007/s00405-016-4285-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2016] [Accepted: 08/24/2016] [Indexed: 11/24/2022]
Abstract
Most trained musicians are actively involved in rigorous practice from several years to achieve a high level of proficiency. Therefore, musicians are best group to research changes or modification in brain structures and functions across several information processing systems. This study aimed to investigate cortical and subcortical processing of short duration speech stimuli in trained rock musicians and non-musicians. Two groups of participant (experimental and control groups) in the age range of 18-25 years were selected for the study. Experimental group includes 15 rock musicians who had minimum professional training of 5 years of rock music, and each member had to be a regular performer of rock music for at least 15 h a week. Further age-matched 15 participants who were not having any formal training of any music served as non-musicians, in the control group. The speech-evoked ABR (S-ABR) and speech-evoked ALLR (S-LLR) with short duration speech 'synthetic /da/' was elicited in both groups. Different measures were analyzed for S-ABR and S-LLR. For S-ABR, MANOVA revealed significant main effect of groups on latencies of wave V, wave A, and amplitude of wave V/A slope. Similarly, Kruskal-Wallis test showed significantly higher F 0 amplitude in rock musicians compared with non-musicians. For S-LLR, MANOVA showed statistically significant differences observed for latencies of wave P2 and N2 and amplitude measures of P2-N2 amplitude. This study indicated better neural processing of short duration speech stimuli at subcortical as well as cortical level among rock musicians when compared with non-musicians.
Collapse
Affiliation(s)
- Prawin Kumar
- All India Institute of Speech and Hearing, Mysuru, 570006, Karnataka, India
| | - Sam Publius Anil
- All India Institute of Speech and Hearing, Mysuru, 570006, Karnataka, India
| | - Vibhu Grover
- All India Institute of Speech and Hearing, Mysuru, 570006, Karnataka, India
| | | | | |
Collapse
|
38
|
Sanju HK, Kumar P. Enhanced auditory evoked potentials in musicians: A review of recent findings. J Otol 2016; 11:63-72. [PMID: 29937812 PMCID: PMC6002589 DOI: 10.1016/j.joto.2016.04.002] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2016] [Revised: 04/25/2016] [Accepted: 04/25/2016] [Indexed: 11/26/2022] Open
Abstract
Auditory evoked potentials serve as an objective mode for assessment to check the functioning of the auditory system and neuroplasticity. Literature has reported enhanced electrophysiological responses in musicians, which shows neuroplasticity in musicians. Various databases including PubMed, Google, Google Scholar and Medline were searched for references related to auditory evoked potentials in musicians from 1994 till date. Different auditory evoked potentials in musicians have been summarized in the present article. The findings of various studies may support as evidences for music-induced neuroplasticity which can be used for the treatment of various clinical disorders. The search results showed enhanced auditory evoked potentials in musicians compared to non-musicians from brainstem to cortical levels. Also, the present review showed enhanced attentive and pre-attentive skills in musicians compared to non-musicians.
Collapse
Affiliation(s)
| | - Prawin Kumar
- Department of Audiology, All India Institute of Speech and Hearing, Mysore, Karnataka, India
| |
Collapse
|
39
|
The Effects of Training on Music Perception and Appreciation for Cochlear Implant Recipients. ACTA ACUST UNITED AC 2016. [DOI: 10.1155/2016/6286514] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/13/2023]
Abstract
The aim of this study was to compare the effectiveness of a music appreciation training program (MATP) to that of focused music listening (FML) for improving music and/or speech in noise perception for postlingually deafened cochlear implant (CI) recipients. It was hypothesized that the MATP would show greater improvements than FML. Ten CI recipients were randomly divided into two groups: one undertaking the MATP and the other undertaking FML. Participants completed four 30-minute sessions per week for 8 weeks, with tests of music and speech-in-noise perception being administered four times per participant: before and after a control period, immediately after the intervention, and 4–8 weeks after intervention. There was a significant pre- to posttraining difference for the MATP group on the instrument identification test, as well as for half of the quality rating assessments. Although no statistically significant improvements were obtained for the FML group, there was a trend of higher scores postintervention for the instrument and ensemble identification tests, and compliance was substantially better than for the MATP group. While the results showed that only the music training significantly improved music perception, the potential of FML to benefit some CI recipients for some tasks was also observed.
Collapse
|
40
|
Abstract
Every day we communicate using complex linguistic and musical systems, yet these modern systems are the product of a much more ancient relationship with sound. When we speak, we communicate not only with the words we choose, but also with the patterns of sound we create and the movements that create them. From the natural rhythms of speech, to the precise timing characteristics of a consonant, these patterns guide our daily communication. By examining the principles of information processing that are common to speech and music, we peel back the layers to reveal the biological foundations of human communication through sound. Further, we consider how the brain's response to sound is shaped by experience, such as musical expertise, and implications for the treatment of communication disorders.
Collapse
Affiliation(s)
- Nina Kraus
- Auditory Neuroscience Laboratory, Departments of
- Communication Sciences,
- Neurobiology and Physiology,
- Otolaryngology, Northwestern University, Evanston, Illinois 60208;
| | - Jessica Slater
- Auditory Neuroscience Laboratory, Departments of
- Communication Sciences,
| |
Collapse
|
41
|
Abstract
UNLABELLED Dyslexics are diagnosed for their poor reading skills, yet they characteristically also suffer from poor verbal memory and often from poor auditory skills. To date, this combined profile has been accounted for in broad cognitive terms. Here we hypothesize that the perceptual deficits associated with dyslexia can be understood computationally as a deficit in integrating prior information with noisy observations. To test this hypothesis we analyzed the performance of human participants in an auditory discrimination task using a two-parameter computational model. One parameter captures the internal noise in representing the current event, and the other captures the impact of recently acquired prior information. Our findings show that dyslexics' perceptual deficit can be accounted for by inadequate adjustment of these components; namely, low weighting of their implicit memory of past trials relative to their internal noise. Underweighting the stimulus statistics decreased dyslexics' ability to compensate for noisy observations. ERP measurements (P2 component) while participants watched a silent movie indicated that dyslexics' perceptual deficiency may stem from poor automatic integration of stimulus statistics. This study provides the first description of a specific computational deficit associated with dyslexia. SIGNIFICANCE STATEMENT This study presents the first attempt to specify the mechanisms underlying dyslexics' perceptual difficulties computationally by applying a specific model, inspired by the Bayesian framework. This model dissociates between the contribution of sensory noise and that of the prior statistics in an auditory perceptual decision task. We show that dyslexics cannot compensate for their perceptual noise by incorporating prior information. By contrast, adequately reading controls' usage of previous information is often close to optimal. We used ERP measurements to assess the neuronal stage of this deficit. We found that unlike their peers, dyslexics' ERP responses are not sensitive to the relations between the current observation and the prior observation, indicating that they cannot establish a reliable prior.
Collapse
|
42
|
Skoe E, Krizman J, Spitzer E, Kraus N. Prior experience biases subcortical sensitivity to sound patterns. J Cogn Neurosci 2015; 27:124-40. [PMID: 25061926 DOI: 10.1162/jocn_a_00691] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/03/2023]
Abstract
To make sense of our ever-changing world, our brains search out patterns. This drive can be so strong that the brain imposes patterns when there are none. The opposite can also occur: The brain can overlook patterns because they do not conform to expectations. In this study, we examined this neural sensitivity to patterns within the auditory brainstem, an evolutionarily ancient part of the brain that can be fine-tuned by experience and is integral to an array of cognitive functions. We have recently shown that this auditory hub is sensitive to patterns embedded within a novel sound stream, and we established a link between neural sensitivity and behavioral indices of learning [Skoe, E., Krizman, J., Spitzer, E., & Kraus, N. The auditory brainstem is a barometer of rapid auditory learning. Neuroscience, 243, 104-114, 2013]. We now ask whether this sensitivity to stimulus statistics is biased by prior experience and the expectations arising from this experience. To address this question, we recorded complex auditory brainstem responses (cABRs) to two patterned sound sequences formed from a set of eight repeating tones. For both patterned sequences, the eight tones were presented such that the transitional probability (TP) between neighboring tones was either 33% (low predictability) or 100% (high predictability). Although both sequences were novel to the healthy young adult listener and had similar TP distributions, one was perceived to be more musical than the other. For the more musical sequence, participants performed above chance when tested on their recognition of the most predictable two-tone combinations within the sequence (TP of 100%); in this case, the cABR differed from a baseline condition where the sound sequence had no predictable structure. In contrast, for the less musical sequence, learning was at chance, suggesting that listeners were "deaf" to the highly predictable repeating two-tone combinations in the sequence. For this condition, the cABR also did not differ from baseline. From this, we posit that the brainstem acts as a Bayesian sound processor, such that it factors in prior knowledge about the environment to index the probability of particular events within ever-changing sensory conditions.
Collapse
|
43
|
Butler BE, Trainor LJ. The Musician Redefined: A Behavioral Assessment of Rhythm Perception in Professional Club DJs. TIMING & TIME PERCEPTION 2015. [DOI: 10.1163/22134468-03002041] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
Studies of musical training demonstrate functional advantages in rhythm tasks that result from enriched auditory experience. Anatomical correlates exist in brain areas involved in auditory perception, speech processing, motor control, attention, and emotion. However, these studies fail to include many classes of musicians that might undergo experience-related change. The current study examined rhythm processing in professional disc jockeys (DJs) who routinely engage in temporally-demanding tasks during practice and performance. In Experiment 1, DJs outperformed controls at detecting a deviation in a rhythmic pattern, and were no different than trained percussionists. In Experiment 2, participants receiving one week of DJ training trended toward outperforming untrained participants on this same measure. Across experiments, movement improved detection of rhythmic deviations, providing evidence of privileged auditory-motor connections, and underscoring the importance of motor areas to rhythm perception. It is clear that DJs show experience-dependent changes in perception that are comparable to more traditional musicians.
Collapse
Affiliation(s)
| | - Laurel J. Trainor
- McMaster UniversityCanada
- McMaster Institute for Music and the MindCanada
- Rotman Research InstituteCanada
| |
Collapse
|
44
|
Smayda KE, Chandrasekaran B, Maddox WT. Enhanced cognitive and perceptual processing: a computational basis for the musician advantage in speech learning. Front Psychol 2015; 6:682. [PMID: 26052304 PMCID: PMC4439769 DOI: 10.3389/fpsyg.2015.00682] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2014] [Accepted: 05/10/2015] [Indexed: 01/05/2023] Open
Abstract
Long-term music training can positively impact speech processing. A recent framework developed to explain such cross-domain plasticity posits that music training-related advantages in speech processing are due to shared cognitive and perceptual processes between music and speech. Although perceptual and cognitive processing advantages due to music training have been independently demonstrated, to date no study has examined perceptual and cognitive processing within the context of a single task. The present study examines the impact of long-term music training on speech learning from a rigorous, computational perspective derived from signal detection theory. Our computational models provide independent estimates of cognitive and perceptual processing in native English-speaking musicians (n = 15, mean age = 25 years) and non-musicians (n = 15, mean age = 23 years) learning to categorize non-native lexical pitch patterns (Mandarin tones). Musicians outperformed non-musicians in this task. Model-based analyses suggested that musicians shifted from simple unidimensional decision strategies to more optimal multidimensional (MD) decision strategies sooner than non-musicians. In addition, musicians used optimal decisional strategies more often than non-musicians. However, musicians and non-musicians who used MD strategies showed no difference in performance. We estimated parameters that quantify the magnitude of perceptual variability along two dimensions that are critical for tone categorization: pitch height and pitch direction. Both musicians and non-musicians showed a decrease in perceptual variability along the pitch height dimension, but only musicians showed a significant reduction in perceptual variability along the pitch direction dimension. Notably, these advantages persisted during a generalization phase, when no feedback was provided. These results provide an insight into the mechanisms underlying the musician advantage observed in non-native speech learning.
Collapse
Affiliation(s)
- Kirsten E Smayda
- Department of Psychology, The University of Texas at Austin Austin, TX USA
| | - Bharath Chandrasekaran
- Department of Psychology, The University of Texas at Austin Austin, TX USA ; Department of Communication Sciences and Disorders, The University of Texas at Austin Austin, TX USA
| | - W Todd Maddox
- Department of Psychology, The University of Texas at Austin Austin, TX USA
| |
Collapse
|
45
|
Pitch expertise is not created equal: Cross-domain effects of musicianship and tone language experience on neural and behavioural discrimination of speech and music. Neuropsychologia 2015; 71:52-63. [DOI: 10.1016/j.neuropsychologia.2015.03.019] [Citation(s) in RCA: 37] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2014] [Revised: 03/07/2015] [Accepted: 03/18/2015] [Indexed: 01/01/2023]
|
46
|
Carey D, Rosen S, Krishnan S, Pearce MT, Shepherd A, Aydelott J, Dick F. Generality and specificity in the effects of musical expertise on perception and cognition. Cognition 2015; 137:81-105. [DOI: 10.1016/j.cognition.2014.12.005] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2013] [Revised: 11/03/2014] [Accepted: 12/18/2014] [Indexed: 10/24/2022]
|
47
|
Boebinger D, Evans S, Rosen S, Lima CF, Manly T, Scott SK. Musicians and non-musicians are equally adept at perceiving masked speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 137:378-87. [PMID: 25618067 PMCID: PMC4434218 DOI: 10.1121/1.4904537] [Citation(s) in RCA: 98] [Impact Index Per Article: 10.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/19/2023]
Abstract
There is much interest in the idea that musicians perform better than non-musicians in understanding speech in background noise. Research in this area has often used energetic maskers, which have their effects primarily at the auditory periphery. However, masking interference can also occur at more central auditory levels, known as informational masking. This experiment extends existing research by using multiple maskers that vary in their informational content and similarity to speech, in order to examine differences in perception of masked speech between trained musicians (n = 25) and non-musicians (n = 25). Although musicians outperformed non-musicians on a measure of frequency discrimination, they showed no advantage in perceiving masked speech. Further analysis revealed that non-verbal IQ, rather than musicianship, significantly predicted speech reception thresholds in noise. The results strongly suggest that the contribution of general cognitive abilities needs to be taken into account in any investigations of individual variability for perceiving speech in noise.
Collapse
Affiliation(s)
- Dana Boebinger
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Samuel Evans
- Institute of Cognitive Neuroscience, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom
| | - Stuart Rosen
- Speech, Hearing & Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 2PF, United Kingdom
| | - César F Lima
- Centre for Psychology at University of Porto, Rua Alfredo Allen, 4200-135 Porto, Portugal
| | - Tom Manly
- Medical Research Council Cognition and Brain Sciences Unit, Cambridge, 15 Chaucer Road, Cambridge CB2 7EF, United Kingdom
| | - Sophie K Scott
- Division of Psychology and Language Sciences, University College London, Gower Street, London WC1E 6BT, United Kingdom
| |
Collapse
|
48
|
Kraus N, Slater J. Music and language. THE HUMAN AUDITORY SYSTEM - FUNDAMENTAL ORGANIZATION AND CLINICAL DISORDERS 2015; 129:207-22. [DOI: 10.1016/b978-0-444-62630-1.00012-3] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/10/2023]
|
49
|
Dopamine receptor D4 (DRD4) gene modulates the influence of informational masking on speech recognition. Neuropsychologia 2014; 67:121-31. [PMID: 25497692 DOI: 10.1016/j.neuropsychologia.2014.12.013] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2014] [Revised: 12/09/2014] [Accepted: 12/10/2014] [Indexed: 12/30/2022]
Abstract
Listeners vary substantially in their ability to recognize speech in noisy environments. Here we examined the role of genetic variation on individual differences in speech recognition in various noise backgrounds. Background noise typically varies in the levels of energetic masking (EM) and informational masking (IM) imposed on target speech. Relative to EM, release from IM is hypothesized to place greater demand on executive function to selectively attend to target speech while ignoring competing noises. Recent evidence suggests that the long allele variant in exon III of the DRD4 gene, primarily expressed in the prefrontal cortex, may be associated with enhanced selective attention to goal-relevant high-priority information even in the face of interference. We investigated the extent to which this polymorphism is associated with speech recognition in IM and EM conditions. In an unscreened adult sample (Experiment 1) and a larger screened replication sample (Experiment 2), we demonstrate that individuals with the DRD4 long variant show better recognition performance in noise conditions involving significant IM, but not in EM conditions. In Experiment 2, we also obtained neuropsychological measures to assess the underlying mechanisms. Mediation analysis revealed that this listening condition-specific advantage was mediated by enhanced executive attention/working memory capacity in individuals with the long allele variant. These findings suggest that DRD4 may contribute specifically to individual differences in speech recognition ability in noise conditions that place demands on executive function.
Collapse
|
50
|
Zendel BR, Tremblay CD, Belleville S, Peretz I. The impact of musicianship on the cortical mechanisms related to separating speech from background noise. J Cogn Neurosci 2014; 27:1044-59. [PMID: 25390195 DOI: 10.1162/jocn_a_00758] [Citation(s) in RCA: 37] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Musicians have enhanced auditory processing abilities. In some studies, these abilities are paralleled by an improved understanding of speech in noisy environments, partially due to more robust encoding of speech signals in noise at the level of the brainstem. Little is known about the impact of musicianship on attention-dependent cortical activity related to lexical access during a speech-in-noise task. To address this issue, we presented musicians and nonmusicians with single words mixed with three levels of background noise, across two conditions, while monitoring electrical brain activity. In the active condition, listeners repeated the words aloud, and in the passive condition, they ignored the words and watched a silent film. When background noise was most intense, musicians repeated more words correctly compared with nonmusicians. Auditory evoked responses were attenuated and delayed with the addition of background noise. In musicians, P1 amplitude was marginally enhanced during active listening and was related to task performance in the most difficult listening condition. By comparing ERPs from the active and passive conditions, we isolated an N400 related to lexical access. The amplitude of the N400 was not influenced by the level of background noise in musicians, whereas N400 amplitude increased with the level of background noise in nonmusicians. In nonmusicians, the increase in N400 amplitude was related to a reduction in task performance. In musicians only, there was a rightward shift of the sources contributing to the N400 as the level of background noise increased. This pattern of results supports the hypothesis that encoding of speech in noise is more robust in musicians and suggests that this facilitates lexical access. Moreover, the shift in sources suggests that musicians, to a greater extent than nonmusicians, may increasingly rely on acoustic cues to understand speech in noise.
Collapse
Affiliation(s)
- Benjamin Rich Zendel
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montréal, Québec, Canada
| | | | | | | |
Collapse
|