1
|
Fallahnezhad T, Pourbakht A, Toufan R, Jalaei S. The Effect of Combined Auditory Training on Concurrent Sound Segregation in Young old: A Single-Blinded Randomized Clinical Trial. Indian J Otolaryngol Head Neck Surg 2023:1-7. [PMID: 37362117 PMCID: PMC10236386 DOI: 10.1007/s12070-023-03923-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Accepted: 05/29/2023] [Indexed: 06/28/2023] Open
Abstract
This study aimed to investigate the behavioral results of perceptual learning in young old using double-vowel discrimination tasks in combined auditory training programs. In a single-blind randomized clinical trial, 35 participants were randomly divided into three groups and received different auditory training programs for six sessions using the software. To compare the double-vowel discrimination score, CV in noise test, and reaction time to the first and second vowels pre- and post-intervention, an analysis of variance was conducted. The discrimination score in the double vowel task and CV in noise test improved after training with no significant difference between the groups. After auditory training, the lowest RT1 was observed in the first intervention group, whereas RT2 decreased only in the second intervention. The present study showed that combined auditory training programs are as effective as conventional auditory training programs in improving speech perception in the elderly. Modifications in the sensory cortex could be investigated using electrophysiological recordings, but this was not conducted because of the pandemic. Supplementary Information The online version contains supplementary material available at 10.1007/s12070-023-03923-x.
Collapse
Affiliation(s)
- Tayyebe Fallahnezhad
- Rehabilitation research center, Department of Audiology, School of Rehabilitation Sciences, University of Medical Sciences, Tehran, Iran
| | - Akram Pourbakht
- Rehabilitation research center, Department of Audiology, School of Rehabilitation Sciences, University of Medical Sciences, Tehran, Iran
| | - Reyhane Toufan
- Rehabilitation research center, Department of Audiology, School of Rehabilitation Sciences, University of Medical Sciences, Tehran, Iran
| | - Shohre Jalaei
- Department of Physiotherapy, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
2
|
Gohari N, Hosseini Dastgerdi Z, Bernstein LJ, Alain C. Neural correlates of concurrent sound perception: A review and guidelines for future research. Brain Cogn 2022; 163:105914. [PMID: 36155348 DOI: 10.1016/j.bandc.2022.105914] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2022] [Revised: 08/30/2022] [Accepted: 09/02/2022] [Indexed: 11/02/2022]
Abstract
The perception of concurrent sound sources depends on processes (i.e., auditory scene analysis) that fuse and segregate acoustic features according to harmonic relations, temporal coherence, and binaural cues (encompass dichotic pitch, location difference, simulated echo). The object-related negativity (ORN) and P400 are electrophysiological indices of concurrent sound perception. Here, we review the different paradigms used to study concurrent sound perception and the brain responses obtained from these paradigms. Recommendations regarding the design and recording parameters of the ORN and P400 are made, and their clinical applications in assessing central auditory processing ability in different populations are discussed.
Collapse
Affiliation(s)
- Nasrin Gohari
- Department of Audiology, School of Rehabilitation, Hamadan University of Medical Sciences, Hamadan, Iran.
| | - Zahra Hosseini Dastgerdi
- Department of Audiology, School of Rehabilitation, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Lori J Bernstein
- Department of Supportive Care, University Health Network, and Department of Psychiatry, University of Toronto, Toronto, Canada
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre for Geriatric Care & Department of Psychology, University of Toronto, Canada
| |
Collapse
|
3
|
Elmahallawi TH, Gabr TA, Darwish ME, Seleem FM. Children with developmental language disorder: a frequency following response in the noise study. Braz J Otorhinolaryngol 2021; 88:954-961. [PMID: 33766501 PMCID: PMC9615520 DOI: 10.1016/j.bjorl.2021.01.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2020] [Revised: 12/21/2020] [Accepted: 01/31/2021] [Indexed: 11/27/2022] Open
Abstract
Introduction Children with developmental language disorder have been reported to have poor temporal auditory processing. This study aimed to examine the frequency following response. Objective This work aimed to investigate speech processing in quiet and in noise. Methods Two groups of children were included in this work: the control group (15 children with normal language development) and the study group (25 children diagnosed with developmental language disorder). All children were submitted to intelligence scale, language assessment, full audiological evaluation, and frequency following response in quiet and noise (+5QNR and +10QNR). Results Results showed no statically significant difference between both groups as regards IQ or PTA. In the study group, the advanced analysis of frequency following response showed reduced F0 and F2 amplitudes. Results also showed that noise has an impact on both the transient and sustained components of the frequency following response in the same group. Conclusion Children with developmental language disorder have difficulty in speech processing especially in the presence of background noise. Frequency following response is an efficient procedure that can be used to address speech processing problems in children with developmental language disorder.
Collapse
Affiliation(s)
- Trandil H Elmahallawi
- Tanta University Hospitals, Otolaryngology Head and Neck Surgery Department, Audiovestibular Unit, Tanta, Egypt
| | - Takwa A Gabr
- Kafrelsheikh University Hospitals, Otolaryngology Head and Neck Surgery Department, Audiovestibular Unit, Kafrelsheikh, Egypt.
| | - Mohamed E Darwish
- Tanta University Hospitals, Otolaryngology Head and Neck Surgery Department, Phoniatrics Unit, Tanta, Egypt
| | - Fatma M Seleem
- Tanta University Hospitals, Otolaryngology Head and Neck Surgery Department, Audiovestibular Unit, Tanta, Egypt
| |
Collapse
|
4
|
Ross B, Tremblay KL, Alain C. Simultaneous EEG and MEG recordings reveal vocal pitch elicited cortical gamma oscillations in young and older adults. Neuroimage 2019; 204:116253. [PMID: 31600592 DOI: 10.1016/j.neuroimage.2019.116253] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2019] [Revised: 09/13/2019] [Accepted: 10/06/2019] [Indexed: 10/25/2022] Open
Abstract
The frequency-following response with origin in the auditory brainstem represents the pitch contour of voice and can be recorded with electrodes from the scalp. MEG studies also revealed a cortical contribution to the high gamma oscillations at the fundamental frequency (f0) of a vowel stimulus. Therefore, studying the cortical component of the frequency-following response could provide insights into how pitch information is encoded at the cortical level. Comparing how aging affects the different responses may help to uncover the neural mechanisms underlying speech understanding deficits in older age. We simultaneously recorded EEG and MEG responses to the syllable /ba/. MEG beamformer analysis localized sources in bilateral auditory cortices and the midbrain. Time-frequency analysis showed a faithful representation of the pitch contour between 106 Hz and 138 Hz in the cortical activity. A cross-correlation revealed a latency of 20 ms. Furthermore, stimulus onsets elicited cortical 40-Hz responses. Both the 40-Hz and the f0 response amplitudes increased in older age and were larger in the right hemisphere. The effects of aging and laterality of the f0 response were evident in the MEG only, suggesting that both effects were characteristics of the cortical response. After comparing f0 and N1 responses in EEG and MEG, we estimated that approximately one-third of the scalp-recorded f0 response could be cortical in origin. We attributed the significance of the cortical f0 response to the precise timing of cortical neurons that serve as a time-sensitive code for pitch.
Collapse
Affiliation(s)
- Bernhard Ross
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department for Medical Biophysics, University of Toronto, Ontario, Canada.
| | - Kelly L Tremblay
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre, Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Ontario, Canada
| |
Collapse
|
5
|
Brainstem encoding of frequency-modulated sweeps is relevant to Mandarin concurrent-vowels identification for normal-hearing and hearing-impaired listeners. Hear Res 2019; 380:123-136. [DOI: 10.1016/j.heares.2019.06.005] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/05/2018] [Revised: 05/21/2019] [Accepted: 06/25/2019] [Indexed: 11/22/2022]
|
6
|
Yellamsetty A, Bidelman GM. Brainstem correlates of concurrent speech identification in adverse listening conditions. Brain Res 2019; 1714:182-192. [PMID: 30796895 DOI: 10.1016/j.brainres.2019.02.025] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2018] [Revised: 01/07/2019] [Accepted: 02/19/2019] [Indexed: 01/20/2023]
Abstract
When two voices compete, listeners can segregate and identify concurrent speech sounds using pitch (fundamental frequency, F0) and timbre (harmonic) cues. Speech perception is also hindered by the signal-to-noise ratio (SNR). How clear and degraded concurrent speech sounds are represented at early, pre-attentive stages of the auditory system is not well understood. To this end, we measured scalp-recorded frequency-following responses (FFR) from the EEG while human listeners heard two concurrently presented, steady-state (time-invariant) vowels whose F0 differed by zero or four semitones (ST) presented diotically in either clean (no noise) or noise-degraded (+5dB SNR) conditions. Listeners also performed a speeded double vowel identification task in which they were required to identify both vowels correctly. Behavioral results showed that speech identification accuracy increased with F0 differences between vowels, and this perceptual F0 benefit was larger for clean compared to noise degraded (+5dB SNR) stimuli. Neurophysiological data demonstrated more robust FFR F0 amplitudes for single compared to double vowels and considerably weaker responses in noise. F0 amplitudes showed speech-on-speech masking effects, along with a non-linear constructive interference at 0ST, and suppression effects at 4ST. Correlations showed that FFR F0 amplitudes failed to predict listeners' identification accuracy. In contrast, FFR F1 amplitudes were associated with faster reaction times, although this correlation was limited to noise conditions. The limited number of brain-behavior associations suggests subcortical activity mainly reflects exogenous processing rather than perceptual correlates of concurrent speech perception. Collectively, our results demonstrate that FFRs reflect pre-attentive coding of concurrent auditory stimuli that only weakly predict the success of identifying concurrent speech.
Collapse
Affiliation(s)
- Anusha Yellamsetty
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Department of Communication Sciences & Disorders, University of South Florida, USA.
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; University of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
7
|
Stuckenberg MV, Nayak CV, Meyer BT, Völker C, Hohmann V, Bendixen A. Age Effects on Concurrent Speech Segregation by Onset Asynchrony. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:177-189. [PMID: 30534994 DOI: 10.1044/2018_jslhr-h-18-0064] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Purpose For elderly listeners, it is more challenging to listen to 1 voice surrounded by other voices than for young listeners. This could be caused by a reduced ability to use acoustic cues-such as slight differences in onset time-for the segregation of concurrent speech signals. Here, we study whether the ability to benefit from onset asynchrony differs between young (18-33 years) and elderly (55-74 years) listeners. Method We investigated young (normal hearing, N = 20) and elderly (mildly hearing impaired, N = 26) listeners' ability to segregate 2 vowels with onset asynchronies ranging from 20 to 100 ms. Behavioral measures were complemented by a specific event-related brain potential component, the object-related negativity, indicating the perception of 2 distinct auditory objects. Results Elderly listeners' behavioral performance (identification accuracy of the 2 vowels) was considerably poorer than young listeners'. However, both age groups showed the same amount of improvement with increasing onset asynchrony. Object-related negativity amplitude also increased similarly in both age groups. Conclusion Both age groups benefit to a similar extent from onset asynchrony as a cue for concurrent speech segregation during active (behavioral measurement) and during passive (electroencephalographic measurement) listening.
Collapse
Affiliation(s)
- Maria V Stuckenberg
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
- Department of Psychology, University of Leipzig, Germany
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| | - Chaitra V Nayak
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
| | - Bernd T Meyer
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
| | - Christoph Völker
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
| | - Volker Hohmann
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
| | - Alexandra Bendixen
- Cluster of Excellence "Hearing4all," Carl von Ossietzky University of Oldenburg, Germany
- Faculty of Natural Sciences, Chemnitz University of Technology, Germany
| |
Collapse
|
8
|
Zhang C, Tao R, Zhao H. Auditory spatial attention modulates the unmasking effect of perceptual separation in a "cocktail party" environment. Neuropsychologia 2019; 124:108-116. [PMID: 30659864 DOI: 10.1016/j.neuropsychologia.2019.01.009] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2018] [Revised: 11/01/2018] [Accepted: 01/15/2019] [Indexed: 11/30/2022]
Abstract
The perceptual separation between a signal speech and a competing speech (masker), induced by the precedence effect, plays an important role in releasing the signal speech from the masker, especially in a reverberant environment. The perceptual-separation-induced unmasking effect has been suggested to involve multiple cognitive processes, such as selective attention. However, whether listeners' spatial attention modulate the perceptual-separation-induced unmasking effect is not clear. The present study investigated how perceptual separation and auditory spatial attention interact with each other to facilitate speech perception under a simulated noisy and reverberant environment by analyzing the cortical auditory evoked potentials to the signal speech. The results showed that the N1 wave was significantly enhanced by perceptual separation between the signal and masker regardless of whether the participants' spatial attention was directed to the signal or not. However, the P2 wave was significantly enhanced by perceptual separation only when the participants attended to the signal speech. The results indicate that the perceptual-separation-induced facilitation of P2 needs more attentional resource than that of N1. The results also showed that the signal speech caused an enhanced N1 in the contralateral hemisphere regardless of whether participants' attention was directed to the signal or not. In contrast, the signal speech caused an enhanced P2 in the contralateral hemisphere only when the participant attended to the signal. The results indicate that the hemispheric distribution of N1 is mainly affected by the perceptual features of the acoustic stimuli, while that of P2 is affected by the listeners' attentional status.
Collapse
Affiliation(s)
- Changxin Zhang
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China.
| | - Renxia Tao
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China
| | - Hang Zhao
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China
| |
Collapse
|
9
|
Abstract
Cochlear implants restore hearing in deaf individuals, but speech perception remains challenging. Poor discrimination of spectral components is thought to account for limitations of speech recognition in cochlear implant users. We investigated how combined variations of spectral components along two orthogonal dimensions can maximize neural discrimination between two vowels, as measured by mismatch negativity. Adult cochlear implant users and matched normal-hearing listeners underwent electroencephalographic event-related potentials recordings in an optimum-1 oddball paradigm. A standard /a/ vowel was delivered in an acoustic free field along with stimuli having a deviant fundamental frequency (+3 and +6 semitones), a deviant first formant making it a /i/ vowel or combined deviant fundamental frequency and first formant (+3 and +6 semitones /i/ vowels). Speech recognition was assessed with a word repetition task. An analysis of variance between both amplitude and latency of mismatch negativity elicited by each deviant vowel was performed. The strength of correlations between these parameters of mismatch negativity and speech recognition as well as participants' age was assessed. Amplitude of mismatch negativity was weaker in cochlear implant users but was maximized by variations of vowels' first formant. Latency of mismatch negativity was later in cochlear implant users and was particularly extended by variations of the fundamental frequency. Speech recognition correlated with parameters of mismatch negativity elicited by the specific variation of the first formant. This nonlinear effect of acoustic parameters on neural discrimination of vowels has implications for implant processor programming and aural rehabilitation.
Collapse
Affiliation(s)
- François Prévost
- 1 Department of Speech Pathology and Audiology, McGill University Health Centre, Montreal, Quebec, Canada.,2 International Laboratory for Brain, Music & Sound Research, Montreal, Quebec, Canada
| | - Alexandre Lehmann
- 2 International Laboratory for Brain, Music & Sound Research, Montreal, Quebec, Canada.,3 Department of Otolaryngology-Head and Neck Surgery, McGill University, Montreal, Quebec, Canada.,4 Centre for Research on Brain, Language & Music, Montreal, Quebec, Canada
| |
Collapse
|
10
|
Gandras K, Grimm S, Bendixen A. Electrophysiological Correlates of Speaker Segregation and Foreground-Background Selection in Ambiguous Listening Situations. Neuroscience 2018; 389:19-29. [PMID: 28735101 DOI: 10.1016/j.neuroscience.2017.07.021] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2017] [Revised: 07/10/2017] [Accepted: 07/10/2017] [Indexed: 11/15/2022]
Abstract
In everyday listening environments, a main task for our auditory system is to follow one out of multiple speakers talking simultaneously. The present study was designed to find electrophysiological indicators of two central processes involved - segregating the speech mixture into distinct speech sequences corresponding to the two speakers, and then attending to one of the speech sequences. We generated multistable speech stimuli that were set up to create ambiguity as to whether only one or two speakers are talking. Thereby we were able to investigate three perceptual alternatives (no segregation, segregated - speaker A in the foreground, segregated - speaker B in the foreground) without any confounding stimulus changes. Participants listened to a continuously repeating sequence of syllables, which were uttered alternately by two human speakers, and indicated whether they perceived the sequence as an inseparable mixture or as originating from two separate speakers. In the latter case, they distinguished which speaker was in their attentional foreground. Our data show a long-lasting event-related potential (ERP) modulation starting at 130ms after stimulus onset, which can be explained by the perceptual organization of the two speech sequences into attended foreground and ignored background streams. Our paradigm extends previous work with pure-tone sequences toward speech stimuli and adds the possibility to obtain neural correlates of the difficulty to segregate a speech mixture into distinct streams.
Collapse
Affiliation(s)
- Katharina Gandras
- Department of Psychology, Cluster of Excellence "Hearing4all", European Medical School, Carl von Ossietzky University of Oldenburg, D-26111 Oldenburg, Germany.
| | - Sabine Grimm
- Department of Physics, School of Natural Sciences, Chemnitz University of Technology, D-09126 Chemnitz, Germany.
| | - Alexandra Bendixen
- Department of Psychology, Cluster of Excellence "Hearing4all", European Medical School, Carl von Ossietzky University of Oldenburg, D-26111 Oldenburg, Germany; Department of Physics, School of Natural Sciences, Chemnitz University of Technology, D-09126 Chemnitz, Germany.
| |
Collapse
|
11
|
Alain C, Du Y, Bernstein LJ, Barten T, Banai K. Listening under difficult conditions: An activation likelihood estimation meta-analysis. Hum Brain Mapp 2018. [PMID: 29536592 DOI: 10.1002/hbm.24031] [Citation(s) in RCA: 73] [Impact Index Per Article: 12.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023] Open
Abstract
The brain networks supporting speech identification and comprehension under difficult listening conditions are not well specified. The networks hypothesized to underlie effortful listening include regions responsible for executive control. We conducted meta-analyses of auditory neuroimaging studies to determine whether a common activation pattern of the frontal lobe supports effortful listening under different speech manipulations. Fifty-three functional neuroimaging studies investigating speech perception were divided into three independent Activation Likelihood Estimate analyses based on the type of speech manipulation paradigm used: Speech-in-noise (SIN, 16 studies, involving 224 participants); spectrally degraded speech using filtering techniques (15 studies involving 270 participants); and linguistic complexity (i.e., levels of syntactic, lexical and semantic intricacy/density, 22 studies, involving 348 participants). Meta-analysis of the SIN studies revealed higher effort was associated with activation in left inferior frontal gyrus (IFG), left inferior parietal lobule, and right insula. Studies using spectrally degraded speech demonstrated increased activation of the insula bilaterally and the left superior temporal gyrus (STG). Studies manipulating linguistic complexity showed activation in the left IFG, right middle frontal gyrus, left middle temporal gyrus and bilateral STG. Planned contrasts revealed left IFG activation in linguistic complexity studies, which differed from activation patterns observed in SIN or spectral degradation studies. Although there were no significant overlap in prefrontal activation across these three speech manipulation paradigms, SIN and spectral degradation showed overlapping regions in left and right insula. These findings provide evidence that there is regional specialization within the left IFG and differential executive networks underlie effortful listening.
Collapse
Affiliation(s)
- Claude Alain
- Rotman Research Institute, Baycrest Health Centre, Toronto, Ontario, Canada.,Department of Psychology, University of Toronto, Toronto, Ontario, Canada
| | - Yi Du
- CAS Key Laboratory of Behavioral Science, Institute of Psychology, Chinese Academy of Sciences, Beijing, China
| | - Lori J Bernstein
- Department of Supportive Care, Princess Margaret Cancer Centre, University Health Network, Toronto, Ontario, Canada.,Department of Psychiatry, University of Toronto, Toronto, Ontario, Canada
| | - Thijs Barten
- Rotman Research Institute, Baycrest Health Centre, Toronto, Ontario, Canada
| | - Karen Banai
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
| |
Collapse
|
12
|
Yellamsetty A, Bidelman GM. Low- and high-frequency cortical brain oscillations reflect dissociable mechanisms of concurrent speech segregation in noise. Hear Res 2018; 361:92-102. [PMID: 29398142 DOI: 10.1016/j.heares.2018.01.006] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/24/2017] [Revised: 12/09/2017] [Accepted: 01/12/2018] [Indexed: 10/18/2022]
Abstract
Parsing simultaneous speech requires listeners use pitch-guided segregation which can be affected by the signal-to-noise ratio (SNR) in the auditory scene. The interaction of these two cues may occur at multiple levels within the cortex. The aims of the current study were to assess the correspondence between oscillatory brain rhythms and determine how listeners exploit pitch and SNR cues to successfully segregate concurrent speech. We recorded electrical brain activity while participants heard double-vowel stimuli whose fundamental frequencies (F0s) differed by zero or four semitones (STs) presented in either clean or noise-degraded (+5 dB SNR) conditions. We found that behavioral identification was more accurate for vowel mixtures with larger pitch separations but F0 benefit interacted with noise. Time-frequency analysis decomposed the EEG into different spectrotemporal frequency bands. Low-frequency (θ, β) responses were elevated when speech did not contain pitch cues (0ST > 4ST) or was noisy, suggesting a correlate of increased listening effort and/or memory demands. Contrastively, γ power increments were observed for changes in both pitch (0ST > 4ST) and SNR (clean > noise), suggesting high-frequency bands carry information related to acoustic features and the quality of speech representations. Brain-behavior associations corroborated these effects; modulations in low-frequency rhythms predicted the speed of listeners' perceptual decisions with higher bands predicting identification accuracy. Results are consistent with the notion that neural oscillations reflect both automatic (pre-perceptual) and controlled (post-perceptual) mechanisms of speech processing that are largely divisible into high- and low-frequency bands of human brain rhythms.
Collapse
Affiliation(s)
- Anusha Yellamsetty
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA
| | - Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, USA; Univeristy of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, USA.
| |
Collapse
|
13
|
Bidelman GM, Yellamsetty A. Noise and pitch interact during the cortical segregation of concurrent speech. Hear Res 2017; 351:34-44. [PMID: 28578876 DOI: 10.1016/j.heares.2017.05.008] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/24/2017] [Revised: 05/09/2017] [Accepted: 05/23/2017] [Indexed: 10/19/2022]
Abstract
Behavioral studies reveal listeners exploit intrinsic differences in voice fundamental frequency (F0) to segregate concurrent speech sounds-the so-called "F0-benefit." More favorable signal-to-noise ratio (SNR) in the environment, an extrinsic acoustic factor, similarly benefits the parsing of simultaneous speech. Here, we examined the neurobiological substrates of these two cues in the perceptual segregation of concurrent speech mixtures. We recorded event-related brain potentials (ERPs) while listeners performed a speeded double-vowel identification task. Listeners heard two concurrent vowels whose F0 differed by zero or four semitones presented in either clean (no noise) or noise-degraded (+5 dB SNR) conditions. Behaviorally, listeners were more accurate in correctly identifying both vowels for larger F0 separations but F0-benefit was more pronounced at more favorable SNRs (i.e., pitch × SNR interaction). Analysis of the ERPs revealed that only the P2 wave (∼200 ms) showed a similar F0 x SNR interaction as behavior and was correlated with listeners' perceptual F0-benefit. Neural classifiers applied to the ERPs further suggested that speech sounds are segregated neurally within 200 ms based on SNR whereas segregation based on pitch occurs later in time (400-700 ms). The earlier timing of extrinsic SNR compared to intrinsic F0-based segregation implies that the cortical extraction of speech from noise is more efficient than differentiating speech based on pitch cues alone, which may recruit additional cortical processes. Findings indicate that noise and pitch differences interact relatively early in cerebral cortex and that the brain arrives at the identities of concurrent speech mixtures as early as ∼200 ms.
Collapse
Affiliation(s)
- Gavin M Bidelman
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, 38152, USA; Institute for Intelligent Systems, University of Memphis, Memphis, TN, 38152, USA; Univeristy of Tennessee Health Sciences Center, Department of Anatomy and Neurobiology, Memphis, TN, 38163, USA.
| | - Anusha Yellamsetty
- School of Communication Sciences & Disorders, University of Memphis, Memphis, TN, 38152, USA
| |
Collapse
|
14
|
Kocsis Z, Winkler I, Bendixen A, Alain C. Promoting the perception of two and three concurrent sound objects: An event-related potential study. Int J Psychophysiol 2016; 107:16-28. [PMID: 27374254 DOI: 10.1016/j.ijpsycho.2016.06.016] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2016] [Revised: 06/24/2016] [Accepted: 06/29/2016] [Indexed: 11/16/2022]
Abstract
The auditory environment typically comprises several simultaneously active sound sources. In contrast to the perceptual segregation of two concurrent sounds, the perception of three simultaneous sound objects has not yet been studied systematically. We conducted two experiments in which participants were presented with complex sounds containing sound segregation cues (mistuning, onset asynchrony, differences in frequency or amplitude modulation or in sound location), which were set up to promote the perceptual organization of the tonal elements into one, two, or three concurrent sounds. In Experiment 1, listeners indicated whether they heard one, two, or three concurrent sounds. In Experiment 2, participants watched a silent subtitled movie while EEG was recorded to extract the object-related negativity (ORN) component of the event-related potential. Listeners predominantly reported hearing two sounds when the segregation promoting manipulations were applied to the same tonal element. When two different tonal elements received manipulations promoting them to be heard as separate auditory objects, participants reported hearing two and three concurrent sounds objects with equal probability. The ORN was elicited in most conditions; sounds that included the amplitude- or the frequency-modulation cue generated the smallest ORN amplitudes. Manipulating two different tonal elements yielded numerically and often significantly smaller ORNs than the sum of the ORNs elicited when the same cues were applied on a single tonal element. These results suggest that ORN reflects the presence of multiple concurrent sounds, but not their number. The ORN results are compatible with the horse-race principle of combining different cues of concurrent sound segregation.
Collapse
Affiliation(s)
- Zsuzsanna Kocsis
- Institute of Psychology and Cognitive Neuroscience, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Magyar tudósok körútja 2., Budapest, H-1117, Hungary; Department of Cognitive Science, Faculty of Natural Sciences, Budapest University of Technology and Economics, Egry József u. 1., Budapest, H-1111, Hungary.
| | - István Winkler
- Institute of Psychology and Cognitive Neuroscience, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Magyar tudósok körútja 2., Budapest, H-1117, Hungary; Institute of Psychology, University of Szeged, Egyetem u. 2., Szeged, H-6722, Hungary.
| | - Alexandra Bendixen
- Cognitive Systems Lab, Institute of Physics, Technische Universität Chemnitz, Reichenhainer Str. 70, Chemnitz, D-09126, Germany.
| | - Claude Alain
- Rotman Research Institute, Baycrest Centre, 3560 Bathurst Street, Toronto, Ontario M6A 2E1, Canada; Department of Psychology, University of Toronto, 100 St. George Street, Toronto, Ontario M5S 3G3, Canada.
| |
Collapse
|
15
|
Josupeit A, Kopčo N, Hohmann V. Modeling of speech localization in a multi-talker mixture using periodicity and energy-based auditory features. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 139:2911. [PMID: 27250183 DOI: 10.1121/1.4950699] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
A recent study showed that human listeners are able to localize a short speech target simultaneously masked by four speech tokens in reverberation [Kopčo, Best, and Carlile (2010). J. Acoust. Soc. Am. 127, 1450-1457]. Here, an auditory model for solving this task is introduced. The model has three processing stages: (1) extraction of the instantaneous interaural time difference (ITD) information, (2) selection of target-related ITD information ("glimpses") using a template-matching procedure based on periodicity, spectral energy, or both, and (3) target location estimation. The model performance was compared to the human data, and to the performance of a modified model using an ideal binary mask (IBM) at stage (2). The IBM-based model performed similarly to the subjects, indicating that the binaural model is able to accurately estimate source locations. Template matching using spectral energy and using a combination of spectral energy and periodicity achieved good results, while using periodicity alone led to poor results. Particularly, the glimpses extracted from the initial portion of the signal were critical for good performance. Simulation data show that the auditory features investigated here are sufficient to explain human performance in this challenging listening condition and thus may be used in models of auditory scene analysis.
Collapse
Affiliation(s)
- Angela Josupeit
- Medizinische Physik, Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| | - Norbert Kopčo
- Institute of Computer Science, Faculty of Science, P. J. Šafárik University, Jesenná 5, 04001 Košice, Slovakia
| | - Volker Hohmann
- Medizinische Physik, Cluster of Excellence Hearing4all, Universität Oldenburg, 26111 Oldenburg, Germany
| |
Collapse
|
16
|
Hedrick MS, Moon IJ, Woo J, Won JH. Effects of Physiological Internal Noise on Model Predictions of Concurrent Vowel Identification for Normal-Hearing Listeners. PLoS One 2016; 11:e0149128. [PMID: 26866811 PMCID: PMC4750862 DOI: 10.1371/journal.pone.0149128] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2015] [Accepted: 01/27/2016] [Indexed: 11/18/2022] Open
Abstract
Previous studies have shown that concurrent vowel identification improves with increasing temporal onset asynchrony of the vowels, even if the vowels have the same fundamental frequency. The current study investigated the possible underlying neural processing involved in concurrent vowel perception. The individual vowel stimuli from a previously published study were used as inputs for a phenomenological auditory-nerve (AN) model. Spectrotemporal representations of simulated neural excitation patterns were constructed (i.e., neurograms) and then matched quantitatively with the neurograms of the single vowels using the Neurogram Similarity Index Measure (NSIM). A novel computational decision model was used to predict concurrent vowel identification. To facilitate optimum matches between the model predictions and the behavioral human data, internal noise was added at either neurogram generation or neurogram matching using the NSIM procedure. The best fit to the behavioral data was achieved with a signal-to-noise ratio (SNR) of 8 dB for internal noise added at the neurogram but with a much smaller amount of internal noise (SNR of 60 dB) for internal noise added at the level of the NSIM computations. The results suggest that accurate modeling of concurrent vowel data from listeners with normal hearing may partly depend on internal noise and where internal noise is hypothesized to occur during the concurrent vowel identification process.
Collapse
Affiliation(s)
- Mark S. Hedrick
- Department of Audiology and Speech Pathology, University of Tennessee Health Science Center, Knoxville, TN, United States of America
| | - Il Joon Moon
- Department of Otorhinolaryngology-Head and Neck Surgery, Samsung Medical Center, Sungkyunkwan University, School of Medicine, Seoul, Korea
| | - Jihwan Woo
- Department of Biomedical Engineering, University of Ulsan, Ulsan, Korea
- * E-mail:
| | - Jong Ho Won
- Department of Audiology and Speech Pathology, University of Tennessee Health Science Center, Knoxville, TN, United States of America
| |
Collapse
|
17
|
Vowel perception in listeners with normal hearing and in listeners with hearing loss: a preliminary study. Clin Exp Otorhinolaryngol 2015; 8:26-33. [PMID: 25729492 PMCID: PMC4338088 DOI: 10.3342/ceo.2015.8.1.26] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2013] [Revised: 02/06/2014] [Accepted: 02/18/2014] [Indexed: 11/08/2022] Open
Abstract
OBJECTIVES To determine the influence of hearing loss on perception of vowel slices. METHODS Fourteen listeners aged 20-27 participated; ten (6 males) had hearing within normal limits and four (3 males) had moderate-severe sensorineural hearing loss (SNHL). Stimuli were six naturally-produced words consisting of the vowels /i a u æ ɛ ʌ/ in a /b V b/ context. Each word was presented as a whole and in eight slices: the initial transition, one half and one fourth of initial transition, full central vowel, one-half central vowel, ending transition, one half and one fourth of ending transition. Each of the 54 stimuli was presented 10 times at 70 dB SPL (sound press level); listeners were asked to identify the word. Stimuli were shaped using signal processing software for the listeners with SNHL to mimic gain provided by an appropriately-fitting hearing aid. RESULTS Listeners with SNHL had a steeper rate of decreasing vowel identification with decreasing slice duration as compared to listeners with normal hearing, and the listeners with SNHL showed different patterns of vowel identification across vowels when compared to listeners with normal hearing. CONCLUSION Abnormal temporal integration is likely affecting vowel identification for listeners with SNHL, which in turn affects vowel internal representation at different levels of the auditory system.
Collapse
|
18
|
Sleep-dependent neuroplastic changes during auditory perceptual learning. Neurobiol Learn Mem 2014; 118:133-42. [PMID: 25490057 DOI: 10.1016/j.nlm.2014.12.001] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2014] [Revised: 10/26/2014] [Accepted: 12/02/2014] [Indexed: 11/24/2022]
Abstract
Auditory perceptual learning is accompanied by a significant increase in the amplitude of sensory evoked responses on the second day of training. This is thought to reflect memory consolidation after the first practice session. However, it is unclear whether the changes in sensory evoked responses depend on sleep per se or whether a break between training sessions would sufficiently yield similar changes. To assess the relative contributions of sleep and passage of time (wakefulness) on the sensory evoked responses, we recorded auditory evoked fields using magnetoencephalography while participants performed a vowel segregation task in three different sessions separated by 12h over two consecutive days. The first two practice sessions were scheduled in the morning and evening of the same day for one group and the evening and morning of subsequent days for the other group. For each participant, we modeled the auditory evoked magnetic field with single dipoles in bilateral superior temporal planes. We then examined the amplitudes and latencies of the resulting source waveforms as a function of sleep and passage of time. In both groups, performance gradually improved with repeated testing. Auditory learning was paralleled by increased sustained field between 250 and 350ms after sound onset as well as sensory evoked fields around 200ms after sound onset (i.e., P2m amplitude) for sessions taking place on the same and different days, respectively. These neuromagnetic changes suggest that auditory learning involves a consolidation phase that occurs during the wake state, which is followed by a sleep-dependent consolidation stage indexed by the P2m amplitude.
Collapse
|
19
|
Carey D, Mercure E, Pizzioli F, Aydelott J. Auditory semantic processing in dichotic listening: Effects of competing speech, ear of presentation, and sentential bias on N400s to spoken words in context. Neuropsychologia 2014; 65:102-12. [DOI: 10.1016/j.neuropsychologia.2014.10.016] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2013] [Revised: 08/29/2014] [Accepted: 10/13/2014] [Indexed: 11/16/2022]
|
20
|
Zendel BR, Tremblay CD, Belleville S, Peretz I. The impact of musicianship on the cortical mechanisms related to separating speech from background noise. J Cogn Neurosci 2014; 27:1044-59. [PMID: 25390195 DOI: 10.1162/jocn_a_00758] [Citation(s) in RCA: 37] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Musicians have enhanced auditory processing abilities. In some studies, these abilities are paralleled by an improved understanding of speech in noisy environments, partially due to more robust encoding of speech signals in noise at the level of the brainstem. Little is known about the impact of musicianship on attention-dependent cortical activity related to lexical access during a speech-in-noise task. To address this issue, we presented musicians and nonmusicians with single words mixed with three levels of background noise, across two conditions, while monitoring electrical brain activity. In the active condition, listeners repeated the words aloud, and in the passive condition, they ignored the words and watched a silent film. When background noise was most intense, musicians repeated more words correctly compared with nonmusicians. Auditory evoked responses were attenuated and delayed with the addition of background noise. In musicians, P1 amplitude was marginally enhanced during active listening and was related to task performance in the most difficult listening condition. By comparing ERPs from the active and passive conditions, we isolated an N400 related to lexical access. The amplitude of the N400 was not influenced by the level of background noise in musicians, whereas N400 amplitude increased with the level of background noise in nonmusicians. In nonmusicians, the increase in N400 amplitude was related to a reduction in task performance. In musicians only, there was a rightward shift of the sources contributing to the N400 as the level of background noise increased. This pattern of results supports the hypothesis that encoding of speech in noise is more robust in musicians and suggests that this facilitates lexical access. Moreover, the shift in sources suggests that musicians, to a greater extent than nonmusicians, may increasingly rely on acoustic cues to understand speech in noise.
Collapse
Affiliation(s)
- Benjamin Rich Zendel
- International Laboratory for Brain, Music and Sound Research (BRAMS), Montréal, Québec, Canada
| | | | | | | |
Collapse
|
21
|
Kocsis Z, Winkler I, Szalárdy O, Bendixen A. Effects of multiple congruent cues on concurrent sound segregation during passive and active listening: an event-related potential (ERP) study. Biol Psychol 2014; 100:20-33. [PMID: 24816158 DOI: 10.1016/j.biopsycho.2014.04.005] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2013] [Revised: 04/29/2014] [Accepted: 04/30/2014] [Indexed: 11/17/2022]
Abstract
In two experiments, we assessed the effects of combining different cues of concurrent sound segregation on the object-related negativity (ORN) and the P400 event-related potential components. Participants were presented with sequences of complex tones, half of which contained some manipulation: one or two harmonic partials were mistuned, delayed, or presented from a different location than the rest. In separate conditions, one, two, or three of these manipulations were combined. Participants watched a silent movie (passive listening) or reported after each tone whether they perceived one or two concurrent sounds (active listening). ORN was found in almost all conditions except for location difference alone during passive listening. Combining several cues or manipulating more than one partial consistently led to sub-additive effects on the ORN amplitude. These results support the view that ORN reflects a combined, feature-unspecific assessment of the auditory system regarding the contribution of two sources to the incoming sound.
Collapse
Affiliation(s)
- Zsuzsanna Kocsis
- Institute of Psychology and Cognitive Neuroscience, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Budapest, Hungary; Budapest University of Technology and Economics, Budapest, Hungary.
| | - István Winkler
- Institute of Psychology and Cognitive Neuroscience, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Budapest, Hungary; Institute of Psychology, University of Szeged, Szeged, Hungary
| | - Orsolya Szalárdy
- Institute of Psychology and Cognitive Neuroscience, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Budapest, Hungary
| | - Alexandra Bendixen
- Department of Psychology, Cluster of Excellence "Hearing4all", European Medical School, Carl von Ossietzky University of Oldenburg, Oldenburg, Germany; Department of Psychology, University of Leipzig, Leipzig, Germany
| |
Collapse
|
22
|
Zündorf IC, Lewald J, Karnath HO. Neural correlates of sound localization in complex acoustic environments. PLoS One 2013; 8:e64259. [PMID: 23691185 PMCID: PMC3653868 DOI: 10.1371/journal.pone.0064259] [Citation(s) in RCA: 34] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2013] [Accepted: 04/09/2013] [Indexed: 12/05/2022] Open
Abstract
Listening to and understanding people in a “cocktail-party situation” is a remarkable feature of the human auditory system. Here we investigated the neural correlates of the ability to localize a particular sound among others in an acoustically cluttered environment with healthy subjects. In a sound localization task, five different natural sounds were presented from five virtual spatial locations during functional magnetic resonance imaging (fMRI). Activity related to auditory stream segregation was revealed in posterior superior temporal gyrus bilaterally, anterior insula, supplementary motor area, and frontoparietal network. Moreover, the results indicated critical roles of left planum temporale in extracting the sound of interest among acoustical distracters and the precuneus in orienting spatial attention to the target sound. We hypothesized that the left-sided lateralization of the planum temporale activation is related to the higher specialization of the left hemisphere for analysis of spectrotemporal sound features. Furthermore, the precuneus − a brain area known to be involved in the computation of spatial coordinates across diverse frames of reference for reaching to objects − seems to be also a crucial area for accurately determining locations of auditory targets in an acoustically complex scene of multiple sound sources. The precuneus thus may not only be involved in visuo-motor processes, but may also subserve related functions in the auditory modality.
Collapse
Affiliation(s)
- Ida C. Zündorf
- Division of Neuropsychology, Center of Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
| | - Jörg Lewald
- Department of Cognitive Psychology, Ruhr University Bochum, Bochum, Germany
- Leibniz Research Centre for Working Environment and Human Factors, Dortmund, Germany
| | - Hans-Otto Karnath
- Division of Neuropsychology, Center of Neurology, Hertie Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
- Department of Psychology, University of South Carolina, Columbia, South Carolina, United States of America
- * E-mail:
| |
Collapse
|
23
|
Lee JH. Age-related deficits in the processing of fundamental frequency differences for the intelligibility of competing voices. KOREAN JOURNAL OF AUDIOLOGY 2013; 17:1-8. [PMID: 24653895 PMCID: PMC3936522 DOI: 10.7874/kja.2013.17.1.1] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/26/2012] [Revised: 01/10/2013] [Accepted: 01/13/2013] [Indexed: 11/22/2022]
Abstract
A common complaint of older listeners is that they can hear speech, yet cannot understand it, especially when listening to speech in a background noise. When target and competing speech signals are concurrently presented, a difference in the fundamental frequency (ΔF0) between competing speech signals, which determines the pitch of voice, can be an important and commonly occurring cue to facilitate the separation of the target message from the interfering message, consequently improving intelligibility of the target message. To address the question of whether the older listeners have reduced ability to use ΔF0 and how the age-related deficits in the processing of ΔF0 are theoretically explained, this paper is divided into three parts. The first part of this article summarizes how the speech-communication difficulties that older listeners have are theoretically explained. In the second part, literatures on the perceptual benefits from ΔF0 and the age-related deficits on the use of ΔF0 are reviewed. As a final part, three theoretical models explaining the general processing of ΔF0 are compared to discuss which better explains the age-related deficits in the processing of ΔF0.
Collapse
Affiliation(s)
- Jae Hee Lee
- Department of Audiology and Institute of Audiology, Hallym University of Graduate Studies, Seoul, Korea
| |
Collapse
|
24
|
Alain C, Roye A, Arnott SR. Middle- and long-latency auditory evoked potentials. DISORDERS OF PERIPHERAL AND CENTRAL AUDITORY PROCESSING 2013. [DOI: 10.1016/b978-0-7020-5310-8.00009-0] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/13/2023]
|
25
|
Zendel BR, Alain C. The influence of lifelong musicianship on neurophysiological measures of concurrent sound segregation. J Cogn Neurosci 2012; 25:503-16. [PMID: 23163409 DOI: 10.1162/jocn_a_00329] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
The ability to separate concurrent sounds based on periodicity cues is critical for parsing complex auditory scenes. This ability is enhanced in young adult musicians and reduced in older adults. Here, we investigated the impact of lifelong musicianship on concurrent sound segregation and perception using scalp-recorded ERPs. Older and younger musicians and nonmusicians were presented with periodic harmonic complexes where the second harmonic could be tuned or mistuned by 1-16% of its original value. The likelihood of perceiving two simultaneous sounds increased with mistuning, and musicians, both older and younger, were more likely to detect and report hearing two sounds when the second harmonic was mistuned at or above 2%. The perception of a mistuned harmonic as a separate sound was paralleled by an object-related negativity that was larger and earlier in younger musicians compared with the other three groups. When listeners made a judgment about the harmonic stimuli, the perception of the mistuned harmonic as a separate sound was paralleled by a positive wave at about 400 msec poststimulus (P400), which was enhanced in both older and younger musicians. These findings suggest attention-dependent processing of a mistuned harmonic is enhanced in older musicians and provides further evidence that age-related decline in hearing abilities are mitigated by musical training.
Collapse
|
26
|
Lee JH, Humes LE. Effect of fundamental-frequency and sentence-onset differences on speech-identification performance of young and older adults in a competing-talker background. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 132:1700-17. [PMID: 22978898 PMCID: PMC3460987 DOI: 10.1121/1.4740482] [Citation(s) in RCA: 45] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
This study investigated the benefits of differences between sentences in fundamental frequency (F0) and temporal onset for sentence pairs among listener groups differing in age and hearing sensitivity. Two experiments were completed with the primary difference between experiments being the way in which the stimuli were presented. Experiment 1 used blocked stimulus presentation, which ultimately provided redundant acoustic cues to mark the target sentence in each pair, whereas Experiment 2 sampled a slightly more restricted stimulus space, but in a completely randomized presentation order. For both experiments, listeners were required to detect a cue word ("Baron") for the target sentence in each pair and to then identify the target words (color, number) that appeared later in the target sentence. Results of Experiment 1 showed that F0 or onset separation cues were beneficial to both cue-word detection and color-number identification performance. There were no significant differences across groups in the ability to detect the cue word, but groups differed in their ability to identify the correct color-number words. Elderly adults with impaired hearing had the greatest difficulty with the identification task despite the application of spectral shaping to restore the audibility of the speech stimuli. For the most part, the primary results of Experiment 1 were replicated in Experiment 2, although, in the latter experiment, all older adults, whether they had normal or impaired hearing, performed worse than young adults with normal hearing. From Experiment 2, the benefits received for a difference in F0 between talkers of 6 semitones were equivalent to those received for an onset asynchrony of 300 ms between sentences and, for such conditions, the combination of both sound-segregation cues resulted in an additive benefit.
Collapse
Affiliation(s)
- Jae Hee Lee
- Department of Speech and Hearing Sciences, Indiana University, Bloomington, Indiana 47405, USA.
| | | |
Collapse
|
27
|
Kashino M, Kondo HM. Functional brain networks underlying perceptual switching: auditory streaming and verbal transformations. Philos Trans R Soc Lond B Biol Sci 2012; 367:977-87. [PMID: 22371619 DOI: 10.1098/rstb.2011.0370] [Citation(s) in RCA: 50] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023] Open
Abstract
Recent studies have shown that auditory scene analysis involves distributed neural sites below, in, and beyond the auditory cortex (AC). However, it remains unclear what role each site plays and how they interact in the formation and selection of auditory percepts. We addressed this issue through perceptual multistability phenomena, namely, spontaneous perceptual switching in auditory streaming (AS) for a sequence of repeated triplet tones, and perceptual changes for a repeated word, known as verbal transformations (VTs). An event-related fMRI analysis revealed brain activity timelocked to perceptual switching in the cerebellum for AS, in frontal areas for VT, and the AC and thalamus for both. The results suggest that motor-based prediction, produced by neural networks outside the auditory system, plays essential roles in the segmentation of acoustic sequences both in AS and VT. The frequency of perceptual switching was determined by a balance between the activation of two sites, which are proposed to be involved in exploring novel perceptual organization and stabilizing current perceptual organization. The effect of the gene polymorphism of catechol-O-methyltransferase (COMT) on individual variations in switching frequency suggests that the balance of exploration and stabilization is modulated by catecholamines such as dopamine and noradrenalin. These mechanisms would support the noteworthy flexibility of auditory scene analysis.
Collapse
Affiliation(s)
- Makio Kashino
- NTT Communication Science Laboratories, NTT Corporation, 3-1 Morinosato Wakamiya, Atsugi, Kanagawa 243-0198, Japan.
| | | |
Collapse
|
28
|
Hill KT, Bishop CW, Miller LM. Auditory grouping mechanisms reflect a sound's relative position in a sequence. Front Hum Neurosci 2012; 6:158. [PMID: 22701410 PMCID: PMC3370426 DOI: 10.3389/fnhum.2012.00158] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2012] [Accepted: 05/17/2012] [Indexed: 11/21/2022] Open
Abstract
The human brain uses acoustic cues to decompose complex auditory scenes into its components. For instance to improve communication, a listener can select an individual “stream,” such as a talker in a crowded room, based on cues such as pitch or location. Despite numerous investigations into auditory streaming, few have demonstrated clear correlates of perception; instead, in many studies perception covaries with changes in physical stimulus properties (e.g., frequency separation). In the current report, we employ a classic ABA streaming paradigm and human electroencephalography (EEG) to disentangle the individual contributions of stimulus properties from changes in auditory perception. We find that changes in perceptual state—that is the perception of one versus two auditory streams with physically identical stimuli—and changes in physical stimulus properties are reflected independently in the event-related potential (ERP) during overlapping time windows. These findings emphasize the necessity of controlling for stimulus properties when studying perceptual effects of streaming. Furthermore, the independence of the perceptual effect from stimulus properties suggests the neural correlates of streaming reflect a tone's relative position within a larger sequence (1st, 2nd, 3rd) rather than its acoustics. By clarifying the role of stimulus attributes along with perceptual changes, this study helps explain precisely how the brain is able to distinguish a sound source of interest in an auditory scene.
Collapse
Affiliation(s)
- Kevin T Hill
- Human Neuroimaging Laboratory, Virginia Tech, Virginia Tech Carilion Research Institue, Roanoke VA, USA
| | | | | |
Collapse
|
29
|
Snyder JS, Gregg MK, Weintraub DM, Alain C. Attention, awareness, and the perception of auditory scenes. Front Psychol 2012; 3:15. [PMID: 22347201 PMCID: PMC3273855 DOI: 10.3389/fpsyg.2012.00015] [Citation(s) in RCA: 78] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2011] [Accepted: 01/11/2012] [Indexed: 11/25/2022] Open
Abstract
Auditory perception and cognition entails both low-level and high-level processes, which are likely to interact with each other to create our rich conscious experience of soundscapes. Recent research that we review has revealed numerous influences of high-level factors, such as attention, intention, and prior experience, on conscious auditory perception. And recently, studies have shown that auditory scene analysis tasks can exhibit multistability in a manner very similar to ambiguous visual stimuli, presenting a unique opportunity to study neural correlates of auditory awareness and the extent to which mechanisms of perception are shared across sensory modalities. Research has also led to a growing number of techniques through which auditory perception can be manipulated and even completely suppressed. Such findings have important consequences for our understanding of the mechanisms of perception and also should allow scientists to precisely distinguish the influences of different higher-level influences.
Collapse
Affiliation(s)
- Joel S. Snyder
- Department of Psychology, University of Nevada Las VegasLas Vegas, NV, USA
| | - Melissa K. Gregg
- Department of Psychology, University of Nevada Las VegasLas Vegas, NV, USA
| | - David M. Weintraub
- Department of Psychology, University of Nevada Las VegasLas Vegas, NV, USA
| | - Claude Alain
- The Rotman Research Institute, Baycrest Centre for Geriatric CareToronto, ON, Canada
| |
Collapse
|
30
|
Recording Event-Related Brain Potentials: Application to Study Auditory Perception. THE HUMAN AUDITORY CORTEX 2012. [DOI: 10.1007/978-1-4614-2314-0_4] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
|
31
|
Anderson S, Kraus N. Neural Encoding of Speech and Music: Implications for Hearing Speech in Noise. Semin Hear 2011; 32:129-141. [PMID: 24748717 PMCID: PMC3989107 DOI: 10.1055/s-0031-1277234] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/17/2022] Open
Abstract
Understanding speech in a background of competing noise is challenging, especially for individuals with hearing loss or deficits in auditory processing ability. The ability to hear in background noise cannot be predicted from the audiogram, an assessment of peripheral hearing ability; therefore, it is important to consider the impact of central and cognitive factors on speech-in-noise perception. Auditory processing in complex environments is reflected in neural encoding of pitch, timing, and timbre, the crucial elements of speech and music. Musical expertise in processing pitch, timing, and timbre may transfer to enhancements in speech-in-noise perception due to shared neural pathways for speech and music. Through cognitive-sensory interactions, musicians develop skills enabling them to selectively listen to relevant signals embedded in a network of melodies and harmonies, and this experience leads in turn to enhanced ability to focus on one voice in a background of other voices. Here we review recent work examining the biological mechanisms of speech and music perception and the potential for musical experience to ameliorate speech-in-noise listening difficulties.
Collapse
Affiliation(s)
- Samira Anderson
- Auditory Neuroscience Laboratory, Northwestern University, Evanston, Illinois
- Department of Communication Sciences, Northwestern University, Evanston, Illinois
| | - Nina Kraus
- Auditory Neuroscience Laboratory, Northwestern University, Evanston, Illinois
- Department of Communication Sciences, Northwestern University, Evanston, Illinois
- Department of Neurobiology and Physiology, Northwestern University, Evanston, Illinois
- Department of Otolaryngology, Northwestern University, Evanston, Illinois
| |
Collapse
|
32
|
Arehart KH, Souza PE, Muralimanohar RK, Miller CW. Effects of age on concurrent vowel perception in acoustic and simulated electroacoustic hearing. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2011; 54:190-210. [PMID: 20689036 PMCID: PMC3258509 DOI: 10.1044/1092-4388(2010/09-0145)] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
PURPOSE In this study, the authors investigated the effects of age on the use of fundamental frequency differences (ΔF(0)) in the perception of competing synthesized vowels in simulations of electroacoustic and cochlear-implant hearing. METHOD Twelve younger listeners with normal hearing and 13 older listeners with (near) normal hearing were evaluated in their use of ΔF(0) in the perception of competing synthesized vowels for 3 conditions: unprocessed synthesized vowels (UNP), envelope-vocoded synthesized vowels that simulated a cochlear implant (VOC), and synthesized vowels processed to simulate electroacoustic stimulation (EAS) hearing. Tasks included (a) multiplicity, which required listeners to identify whether a stimulus contained 1 or 2 sounds and (b) double-vowel identification, which required listeners to attach phonemic labels to the competing synthesized vowels. RESULTS Multiplicity perception was facilitated by ΔF(0) in UNP and EAS but not in VOC, with no age-related deficits evident. Double-vowel identification was facilitated by ΔF(0), with ΔF(0) benefit largest in UNP, reduced in EAS, and absent in VOC. Age adversely affected overall identification and ΔF(0) benefit on the double-vowel task. CONCLUSIONS Some but not all older listeners derived ΔF(0) benefit in EAS hearing. This variability may partly be due to how listeners are able to draw on higher-level processing resources in extracting and integrating cues in EAS hearing.
Collapse
Affiliation(s)
- Kathryn H Arehart
- University of Colorado at Boulder - Dept. SLHS, Campus Box 409, Boulder, Colorado 80309, USA.
| | | | | | | |
Collapse
|
33
|
Anderson S, Skoe E, Chandrasekaran B, Zecker S, Kraus N. Brainstem correlates of speech-in-noise perception in children. Hear Res 2010; 270:151-7. [PMID: 20708671 PMCID: PMC2997182 DOI: 10.1016/j.heares.2010.08.001] [Citation(s) in RCA: 63] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/01/2010] [Revised: 07/30/2010] [Accepted: 08/04/2010] [Indexed: 10/19/2022]
Abstract
Children often have difficulty understanding speech in challenging listening environments. In the absence of peripheral hearing loss, these speech perception difficulties may arise from dysfunction at more central levels in the auditory system, including subcortical structures. We examined brainstem encoding of pitch in a speech syllable in 38 school-age children. In children with poor speech-in-noise perception, we find impaired encoding of the fundamental frequency and the second harmonic, two important cues for pitch perception. Pitch, an essential factor in speaker identification, aids the listener in tracking a specific voice from a background of voices. These results suggest that the robustness of subcortical neural encoding of pitch features in time-varying signals is a key factor in determining success with perceiving speech in noise.
Collapse
Affiliation(s)
- Samira Anderson
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL 60208, USA.
| | | | | | | | | |
Collapse
|
34
|
Du Y, He Y, Ross B, Bardouille T, Wu X, Li L, Alain C. Human Auditory Cortex Activity Shows Additive Effects of Spectral and Spatial Cues during Speech Segregation. Cereb Cortex 2010; 21:698-707. [PMID: 20685854 DOI: 10.1093/cercor/bhq136] [Citation(s) in RCA: 39] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Affiliation(s)
- Yi Du
- Department of Psychology, Speech and Hearing Research Center, Key Laboratory on Machine Perception (Ministry of Education), Peking University, Beijing, China 100871
| | | | | | | | | | | | | |
Collapse
|
35
|
Shamma SA, Micheyl C. Behind the scenes of auditory perception. Curr Opin Neurobiol 2010; 20:361-6. [PMID: 20456940 DOI: 10.1016/j.conb.2010.03.009] [Citation(s) in RCA: 66] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2010] [Revised: 03/16/2010] [Accepted: 03/29/2010] [Indexed: 11/30/2022]
Abstract
'Auditory scenes' often contain contributions from multiple acoustic sources. These are usually heard as separate auditory 'streams', which can be selectively followed over time. How and where these auditory streams are formed in the auditory system is one of the most fascinating questions facing auditory scientists today. Findings published within the past two years indicate that both cortical and subcortical processes contribute to the formation of auditory streams, and they raise important questions concerning the roles of primary and secondary areas of auditory cortex in this phenomenon. In addition, these findings underline the importance of taking into account the relative timing of neural responses, and the influence of selective attention, in the search for neural correlates of the perception of auditory streams.
Collapse
Affiliation(s)
- Shihab A Shamma
- Department of Electrical and Computer Engineering & Institute for Systems Research, University of Maryland College Park, United States.
| | | |
Collapse
|
36
|
Bendixen A, Jones SJ, Klump G, Winkler I. Probability dependence and functional separation of the object-related and mismatch negativity event-related potential components. Neuroimage 2010; 50:285-90. [DOI: 10.1016/j.neuroimage.2009.12.037] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2009] [Revised: 12/04/2009] [Accepted: 12/08/2009] [Indexed: 10/20/2022] Open
|
37
|
Improved segregation of simultaneous talkers differentially affects perceptual and cognitive capacity demands for recognizing speech in competing speech. Atten Percept Psychophys 2010; 72:501-16. [DOI: 10.3758/app.72.2.501] [Citation(s) in RCA: 36] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
38
|
Lipp R, Kitterick P, Summerfield Q, Bailey PJ, Paul-Jordanov I. Concurrent sound segregation based on inharmonicity and onset asynchrony. Neuropsychologia 2010; 48:1417-25. [PMID: 20079754 DOI: 10.1016/j.neuropsychologia.2010.01.009] [Citation(s) in RCA: 29] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2009] [Revised: 01/08/2010] [Accepted: 01/08/2010] [Indexed: 11/19/2022]
Abstract
To explore the neural processes underlying concurrent sound segregation, auditory evoked fields (AEFs) were measured using magnetoencephalography (MEG). To induce the segregation of two auditory objects we manipulated harmonicity and onset synchrony. Participants were presented with complex sounds with (i) all harmonics in-tune (ii) the third harmonic mistuned by 8% of its original value (iii) the onset of the third harmonic delayed by 160 ms compared to the other harmonics. During recording, participants listened to the sounds and performed an auditory localisation task whereas in another session they ignored the sounds and performed a visual localisation task. Active and passive listening was chosen to evaluate the contribution of attention on sound segregation. Both cues - inharmonicity and onset asynchrony - elicited sound segregation, as participants were more likely to report correctly on which side they heard the third harmonic when it was mistuned or delayed compared to being in-tune with all other harmonics. AEF activity associated with concurrent sound segregation was identified over both temporal lobes. We found an early deflection at approximately 75 ms (P75m) after sound onset, probably reflecting an automatic registration of the mistuned harmonic. Subsequent deflections, the object-related negativity (ORNm) and a later displacement (P230m) seem to be more general markers of concurrent sound segregation, as they were elicited by both mistuning and delaying the third harmonic. Results indicate that the ORNm reflects relatively automatic, bottom-up sound segregation processes, whereas the P230m is more sensitive to attention, especially with inharmonicity as the cue for concurrent sound segregation.
Collapse
Affiliation(s)
- Rosa Lipp
- Department of Psychology, Clinical Psychology and Neuropsychology, University of Konstanz, Germany.
| | | | | | | | | |
Collapse
|
39
|
Pitch, harmonicity and concurrent sound segregation: psychoacoustical and neurophysiological findings. Hear Res 2009; 266:36-51. [PMID: 19788920 DOI: 10.1016/j.heares.2009.09.012] [Citation(s) in RCA: 91] [Impact Index Per Article: 6.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/08/2009] [Revised: 09/23/2009] [Accepted: 09/24/2009] [Indexed: 11/18/2022]
Abstract
Harmonic complex tones are a particularly important class of sounds found in both speech and music. Although these sounds contain multiple frequency components, they are usually perceived as a coherent whole, with a pitch corresponding to the fundamental frequency (F0). However, when two or more harmonic sounds occur concurrently, e.g., at a cocktail party or in a symphony, the auditory system must separate harmonics and assign them to their respective F0s so that a coherent and veridical representation of the different sounds sources is formed. Here we review both psychophysical and neurophysiological (single-unit and evoked-potential) findings, which provide some insight into how, and how well, the auditory system accomplishes this task. A survey of computational models designed to estimate multiple F0s and segregate concurrent sources is followed by a review of the empirical literature on the perception and neural coding of concurrent harmonic sounds, including vowels, as well as findings obtained using single complex tones with mistuned harmonics.
Collapse
|
40
|
Vakorin VA, Kovacevic N, McIntosh AR. Exploring transient transfer entropy based on a group-wise ICA decomposition of EEG data. Neuroimage 2009; 49:1593-600. [PMID: 19698792 DOI: 10.1016/j.neuroimage.2009.08.027] [Citation(s) in RCA: 38] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2008] [Revised: 06/17/2009] [Accepted: 08/11/2009] [Indexed: 10/20/2022] Open
Abstract
This paper presents a data-driven pipeline for studying asymmetries in mutual interdependencies between distinct components of EEG signal. Due to volume conductance, estimating coherence between scalp electrodes may lead to spurious results. A group-based independent component analysis (ICA), which is conducted across all subjects and conditions simultaneously, is an alternative representation of the EEG measurements. Within this approach, the extracted components are independent in a global sense while short-lived or transient interdependencies may still be present between the components. In this paper, functional roles of the ICA components are specified through a partial least squares (PLS) analysis of task effects within the time course of the derived components. Functional integration is estimated within the information-theoretic approach using transfer entropy analysis based on asymmetries in mutual interdependencies of reconstructed phase dynamics. A secondary PLS analysis is performed to assess robust task-specific changes in transfer entropy estimates between functionally specific components.
Collapse
|
41
|
Hedrick MS, Madix SG. Effect of vowel identity and onset asynchrony on concurrent vowel identification. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2009; 52:696-705. [PMID: 18952855 DOI: 10.1044/1092-4388(2008/07-0094)] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
PURPOSE The purpose of the current study was to determine the effects of vowel identity and temporal onset asynchrony on identification of vowels overlapped in time. METHOD Fourteen listeners with normal hearing, with a mean age of 24 years, participated. The listeners were asked to identify both of a pair of 200-ms vowels (referred to as double vowels) presented either simultaneously or with a temporal asynchrony ranging from 25 ms to 150 ms in 25-ms steps. The stimuli were synthetic steady-state vowels /i ae u / arranged in seven combinations: /u i/, /ae /, / /, / ae/, /ae i/, / i/, and / u/. RESULTS Listeners' responses revealed that one vowel of a pair was identified correctly more often than the other vowel (known as vowel dominance). Vowel dominance effects were seen for 6 of the 7 vowel pairs, and there was improvement of vowel identification with increasing temporal separation between vowels for 5 of the 7 pairs. Vowel pairs with the vowel // consistently yielded improved identification with increases in temporal asynchrony. DISCUSSION Peripheral masking cannot explain the patterns of results of this study. A more parsimonious explanation may be perceptual anchoring.
Collapse
Affiliation(s)
- Mark S Hedrick
- University of Tennessee, Knoxville, Department of Audiology and Speech Pathology, 578 South Stadium Hall, Knoxville, TN 37996-0740, USA.
| | | |
Collapse
|
42
|
Laufer I, Negishi M, Constable RT. Comparator and non-comparator mechanisms of change detection in the context of speech--an ERP study. Neuroimage 2009; 44:546-62. [PMID: 18938250 PMCID: PMC2643129 DOI: 10.1016/j.neuroimage.2008.09.010] [Citation(s) in RCA: 13] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2008] [Revised: 08/11/2008] [Accepted: 09/09/2008] [Indexed: 11/25/2022] Open
Abstract
Automatic change detection reflects a cognitive memory-based comparison mechanism as well as a sensorial non-comparator mechanism based on differential states of refractoriness. The purpose of this study was to examine whether the comparator mechanism of the mismatch negativity component (MMN) is differentially affected by the lexical status of the deviant. Event-related potential (ERP) data was collected during an "oddball" paradigm designed to elicit the MMN from 15 healthy subjects that were involved in a counting task. Topography pattern analysis and source estimation were utilized to examine the deviance (deviants vs. standards), cognitive (deviants vs. control counterparts) and refractoriness (standards vs. control counterparts) effects elicited by standard-deviant pairs ("deh-day"; "day-deh"; "teh-tay") embedded within "oddball" blocks. Our results showed that when the change was salient regardless of lexical status (i.e., the /e:/ to /eI/ transition) the response tapped the comparator based-mechanism of the MMN which was located in the cuneus/posterior cingulate, reflected sensitivity to the novelty of the auditory object, appeared in the P2 latency range and mainly involved topography modulations. In contrast, when the novelty was low (i.e., the /eI/ to /e:/ transition) an acoustic change complex was elicited which involved strength modulations over the P1/N1 range and implicated the middle temporal gyrus. This result pattern also resembled the one displayed by the non-comparator mechanism. These findings suggest spatially and temporally distinct brain activities of comparator mechanisms of change detection in the context of speech.
Collapse
Affiliation(s)
- Ilan Laufer
- Department of Diagnostic Radiology, Yale University School of Medicine, The Anlyan Center, New Haven, CT 06520-8043, USA.
| | | | | |
Collapse
|
43
|
Abstract
The ability to isolate a single sound source among concurrent sources and reverberant energy is necessary for understanding the auditory world. The precedence effect describes a related experimental finding, that when presented with identical sounds from two locations with a short onset asynchrony (on the order of milliseconds), listeners report a single source with a location dominated by the lead sound. Single-cell recordings in multiple animal models have indicated that there are low-level mechanisms that may contribute to the precedence effect, yet psychophysical studies in humans have provided evidence that top-down cognitive processes have a great deal of influence on the perception of simulated echoes. In the present study, event-related potentials evoked by click pairs at and around listeners' echo thresholds indicate that perception of the lead and lag sound as individual sources elicits a negativity between 100 and 250 msec, previously termed the object-related negativity (ORN). Even for physically identical stimuli, the ORN is evident when listeners report hearing, as compared with not hearing, a second sound source. These results define a neural mechanism related to the conscious perception of multiple auditory objects.
Collapse
|
44
|
Diaconescu AO, Kovacevic N, McIntosh AR. Modality-independent processes in cued motor preparation revealed by cortical potentials. Neuroimage 2008; 42:1255-65. [PMID: 18625564 DOI: 10.1016/j.neuroimage.2008.05.024] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2008] [Revised: 04/30/2008] [Accepted: 05/09/2008] [Indexed: 11/16/2022] Open
|
45
|
Abstract
The auditory system can segregate sounds that overlap in time and frequency, if the sounds differ in acoustic properties such as fundamental frequency (f0). However, the neural mechanisms that underlie this ability are poorly understood. Responses of neurons in the inferior colliculus (IC) of the anesthetized chinchilla were measured. The stimuli were harmonic tones, presented alone (single harmonic tones) and in the presence of a second harmonic tone with a different f0 (double harmonic tones). Responses to single harmonic tones exhibited no stimulus-related temporal pattern, or in some cases, a simple envelope modulated at f0. Responses to double harmonic tones exhibited complex slowly modulated discharge patterns. The discharge pattern varied with the difference in f0 and with characteristic frequency. The discharge pattern also varied with the relative levels of the two tones; complex temporal patterns were observed when levels were equal, but as the level difference increased, the discharge pattern reverted to that associated with single harmonic tones. The results indicated that IC neurons convey information about simultaneous sounds in their temporal discharge patterns and that the patterns are produced by interactions between adjacent components in the spectrum. The representation is "low-resolution," in that it does not convey information about single resolved components from either individual sound.
Collapse
Affiliation(s)
- Donal G Sinex
- Department of Psychology, Utah State University, Logan, UT 84322-2810, USA.
| | | |
Collapse
|
46
|
Alain C. Breaking the wave: effects of attention and learning on concurrent sound perception. Hear Res 2007; 229:225-36. [PMID: 17303355 DOI: 10.1016/j.heares.2007.01.011] [Citation(s) in RCA: 84] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/13/2006] [Revised: 12/06/2006] [Accepted: 01/03/2007] [Indexed: 11/19/2022]
Abstract
The auditory surrounding is often complex with many sound sources active simultaneously. Yet listeners are proficient in breaking apart the composite acoustic wave reaching the ears. This achievement is thought to be the result of bottom-up as well as top-down processes that reflect listeners' experience and knowledge of the auditory environment. Here, specific findings concerning the role of bottom-up and top-down (schema-driven) processes on concurrent sound perception are reviewed, with particular emphasis on studies that have used scalp recording of event-related brain potentials. Findings from several studies indicate that frequency periodicity, upon which concurrent sound perception partly depends, is quickly and automatically registered in primary auditory cortex. Moreover, success in identifying concurrent vowels is accompanied by enhanced neural activity, as revealed by functional magnetic resonance imaging, in thalamus, primary auditory cortex and planum temporale. Lastly, listeners' ability to segregate concurrent vowels improves with training and these neuroplastic changes occur rapidly, demonstrating the flexibility of human speech segregation mechanisms. Together, these studies suggest that the primary auditory cortex and the planum temporale play an important role in concurrent sound perception, and reveal a link between thalamo-cortical activation and the successful separation and identification of speech sounds presented simultaneously.
Collapse
Affiliation(s)
- Claude Alain
- Rotman Research Institute, Baycrest Centre for Geriatric Care, Department of Psychology, University of Toronto, Toronto, ONT, Canada.
| |
Collapse
|
47
|
Snyder JS, Alain C. Age-related changes in neural activity associated with concurrent vowel segregation. ACTA ACUST UNITED AC 2005; 24:492-9. [PMID: 16099361 DOI: 10.1016/j.cogbrainres.2005.03.002] [Citation(s) in RCA: 79] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2004] [Revised: 02/25/2005] [Accepted: 03/01/2005] [Indexed: 11/18/2022]
Abstract
Older adults exhibit degraded speech comprehension in complex sound environments, which may be related to overall age-related declines in low-level sound segregation. This hypothesis was tested by measuring event-related potentials (ERPs) while listeners identified two different vowels presented simultaneously. Older adults were less accurate than young adults at identifying the two vowels, although both groups improved similarly with increasing fundamental frequency differences (Deltaf0) between vowels. Reaction time data showed that older adults took more time to process stimuli, especially those with smaller Deltaf0. A negative ERP wave indexing the automatic registration of Deltaf0 (the object-related negativity) was reduced in older adults. In contrast, young and older adults showed a similar pattern of neural activity indexing attentive processing of Deltaf0. The results suggest that aging affects the ability to automatically segregate speech sounds.
Collapse
Affiliation(s)
- Joel S Snyder
- The Rotman Research Institute, Baycrest Centre for Geriatric Care 3560 Bathurst Street, Toronto, ON, Canada M6A 2E1.
| | | |
Collapse
|