1
|
Takai S, Kanno A, Kawase T, Shirakura M, Suzuki J, Nakasato N, Kawashima R, Katori Y. Possibility of additive effects by the presentation of visual information related to distractor sounds on the contra-sound effects of the N100m responses. Hear Res 2023; 434:108778. [PMID: 37105052 DOI: 10.1016/j.heares.2023.108778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/06/2023] [Revised: 04/13/2023] [Accepted: 04/21/2023] [Indexed: 04/29/2023]
Abstract
Auditory-evoked responses can be affected by different types of contralateral sounds or by attention modulation. The present study examined the additive effects of presenting visual information about contralateral sounds as distractions during dichotic listening tasks on the contralateral effects of N100m responses in the auditory-evoked cortex in 16 subjects (12 males and 4 females). In magnetoencephalography, a tone-burst of 500 ms duration at a frequency of 1000 Hz was played to the left ear at a level of 70 dB as a stimulus to elicit the N100m response, and a movie clip was used as a distractor stimulus under audio-only, visual-only, and audio-visual conditions. Subjects were instructed to pay attention to the left ear and press the response button each time they heard a tone-burst stimulus in their left ear. The results suggest that the presentation of visual information related to the contralateral sound, which worked as a distractor, significantly suppressed the amplitude of the N100m response compared with only the contralateral sound condition. In contrast, the presentation of visual information related to contralateral sound did not affect the latency of the N100m response. These results suggest that the integration of contralateral sounds and related movies may have resulted in a more perceptually loaded stimulus and reduced the intensity of attention to tone-bursts. Our findings suggest that selective attention and saliency mechanisms may have cross-modal effects on other modes of perception.
Collapse
Affiliation(s)
- Shunsuke Takai
- Department of Otolaryngology-Head and Neck Surgery, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan.
| | - Akitake Kanno
- Department of Advanced Spintronics Medical Engineering, Graduate School of Engineering, Tohoku University, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8575, Japan; Department of Epileptology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8575, Japan
| | - Tetsuaki Kawase
- Department of Otolaryngology-Head and Neck Surgery, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan; Laboratory of Rehabilitative Auditory Science, Tohoku University Graduate School of Biomedical Engineering, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan; Department of Audiology, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan
| | - Masayuki Shirakura
- Department of Otolaryngology-Head and Neck Surgery, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan
| | - Jun Suzuki
- Department of Otolaryngology-Head and Neck Surgery, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan
| | - Nobukatsu Nakasato
- Department of Advanced Spintronics Medical Engineering, Graduate School of Engineering, Tohoku University, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8575, Japan; Department of Epileptology, Tohoku University Graduate School of Medicine, 2-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8575, Japan
| | - Ryuta Kawashima
- Institute of Development, Aging and Cancer, Tohoku University, 4-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8575, Japan
| | - Yukio Katori
- Department of Otolaryngology-Head and Neck Surgery, Tohoku University Graduate School of Medicine, 1-1 Seiryo-machi, Aoba-ku, Sendai, Miyagi 980-8574, Japan
| |
Collapse
|
2
|
Chai X, Liu M, Huang T, Wu M, Li J, Zhao X, Yan T, Song Y, Zhang YX. Neurophysiological evidence for goal-oriented modulation of speech perception. Cereb Cortex 2022; 33:3910-3921. [PMID: 35972410 DOI: 10.1093/cercor/bhac315] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 07/20/2022] [Accepted: 07/21/2022] [Indexed: 11/14/2022] Open
Abstract
Speech perception depends on the dynamic interplay of bottom-up and top-down information along a hierarchically organized cortical network. Here, we test, for the first time in the human brain, whether neural processing of attended speech is dynamically modulated by task demand using a context-free discrimination paradigm. Electroencephalographic signals were recorded during 3 parallel experiments that differed only in the phonological feature of discrimination (word, vowel, and lexical tone, respectively). The event-related potentials (ERPs) revealed the task modulation of speech processing at approximately 200 ms (P2) after stimulus onset, probably influencing what phonological information to retain in memory. For the phonological comparison of sequential words, task modulation occurred later at approximately 300 ms (N3 and P3), reflecting the engagement of task-specific cognitive processes. The ERP results were consistent with the changes in delta-theta neural oscillations, suggesting the involvement of cortical tracking of speech envelopes. The study thus provides neurophysiological evidence for goal-oriented modulation of attended speech and calls for speech perception models incorporating limited memory capacity and goal-oriented optimization mechanisms.
Collapse
Affiliation(s)
- Xiaoke Chai
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Min Liu
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Ting Huang
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Meiyun Wu
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Jinhong Li
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Xue Zhao
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Tingting Yan
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Yan Song
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| | - Yu-Xuan Zhang
- State Key Laboratory of Cognitive Neuroscience and Learning, IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing 100875, China
| |
Collapse
|
3
|
Zobel BH, Freyman RL, Sanders LD. Spatial release from informational masking enhances the early cortical representation of speech sounds. AUDITORY PERCEPTION & COGNITION 2022; 5:211-237. [PMID: 36160272 PMCID: PMC9494573 DOI: 10.1080/25742442.2022.2088329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2021] [Accepted: 06/04/2022] [Indexed: 06/16/2023]
Abstract
INTRODUCTION Spatial separation between competing speech streams reduces their confusion (informational masking), improving speech processing under challenging listening conditions. The precise stages of auditory processing involved in this benefit are not fully understood. This study used event-related potentials to examine the processing of target speech under conditions of informational masking and its spatial release. METHODS Participants detected noise-vocoded target speech presented with two-talker noise-vocoded masking speech. In separate conditions, the same set of targets were spatially co-located with maskers to produce informational masking and spatially separated from maskers using a perceptual manipulation to release the informational masking. RESULTS An increase in N1 and P2 amplitude, consistent with cortical auditory evoked potentials, and a later sustained positivity (P300) were observed in response to target onsets only under conditions supporting release from informational masking. At target intensities above masking threshold in both spatial conditions, N1 and P2 latencies were shorter when targets and maskers were perceptually separated. DISCUSSION These results indicate that spatial release from informational masking benefits speech representation beginning in the early stages of auditory perception. Additionally, these results suggest that the auditory evoked potential itself may be heavily dependent upon how information is perceptually organized rather than physically organized.
Collapse
Affiliation(s)
- Benjamin H. Zobel
- Department of Psychological and Brain Sciences, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Richard L. Freyman
- Department of Communication Disorders, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Lisa D. Sanders
- Department of Psychological and Brain Sciences, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| |
Collapse
|
4
|
Yang X, Liu L, Yang P, Ding Y, Wang C, Li L. The Effects of Attention on the Syllable-Induced Prepulse Inhibition of the Startle Reflex and Cortical EEG Responses against Energetic or Informational Masking in Humans. Brain Sci 2022; 12:brainsci12050660. [PMID: 35625046 PMCID: PMC9139428 DOI: 10.3390/brainsci12050660] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2022] [Revised: 05/09/2022] [Accepted: 05/15/2022] [Indexed: 11/29/2022] Open
Abstract
Prepulse inhibition (PPI) is the reduction in the acoustic startle reflex (ASR) when the startling stimulus (pulse) is preceded by a weaker, non-starting stimulus. This can be enhanced by facilitating selective attention to the prepulse against a noise-masking background. On the other hand, the facilitation of selective attention to a target speech can release the target speech from masking, particularly from speech informational masking. It is not clear whether attentional regulation also affects PPI in this kind of auditory masking. This study used a speech syllable as the prepulse to examine whether the masker type and perceptual spatial attention can affect the PPI or the scalp EEG responses to the prepulse in healthy younger-adult humans, and whether the ERPs evoked by the prepulse can predict the PPI intensity of the ASR. The results showed that the speech masker produced a larger masking effect than the noise masker, and the perceptual spatial separation facilitated selective attention to the prepulse, enhancing both the N1 component of the prepulse syllable and the PPI of the ASR, particularly when the masker was speech. In addition, there was no significant correlation between the PPI and ERPs under any of the conditions, but the perceptual separation-induced PPI enhancement and ERP N1P2 peak-to-peak amplitude enhancement were correlated under the speech-masking condition. Thus, the attention-mediated PPI is useful for differentiating noise energetic masking and speech informational masking, and the perceptual separation-induced release of the prepulse from informational masking is more associated with attention-mediated early cortical unmasking processing than with energetic masking. However, the processes for the PPI of the ASR and the cortical responses to the prepulse are mediated by different neural mechanisms.
Collapse
Affiliation(s)
- Xiaoqin Yang
- Collaborative Innovation Center for Brain Disorders, Laboratory of Brain Disorders, Beijing Institute of Brain Disorders, Capital Medical University Ministry of Science and Technology, Beijing 100069, China;
| | - Lei Liu
- Key Laboratory on Machine Perception (Ministry of Education), Beijing Key Laboratory of Behavior and Mental Health, School of Psychological and Cognitive Sciences, Peking University, Beijing 100080, China; (L.L.); (P.Y.); (Y.D.)
| | - Pengcheng Yang
- Key Laboratory on Machine Perception (Ministry of Education), Beijing Key Laboratory of Behavior and Mental Health, School of Psychological and Cognitive Sciences, Peking University, Beijing 100080, China; (L.L.); (P.Y.); (Y.D.)
| | - Yu Ding
- Key Laboratory on Machine Perception (Ministry of Education), Beijing Key Laboratory of Behavior and Mental Health, School of Psychological and Cognitive Sciences, Peking University, Beijing 100080, China; (L.L.); (P.Y.); (Y.D.)
- Division of Sports Science and Physical Education, Tsinghua University, Beijing 100084, China
| | - Changming Wang
- Department of Neurosurgery, Xuanwu Hospital, Capital Medical University, Beijing 100053, China;
| | - Liang Li
- Collaborative Innovation Center for Brain Disorders, Laboratory of Brain Disorders, Beijing Institute of Brain Disorders, Capital Medical University Ministry of Science and Technology, Beijing 100069, China;
- Key Laboratory on Machine Perception (Ministry of Education), Beijing Key Laboratory of Behavior and Mental Health, School of Psychological and Cognitive Sciences, Peking University, Beijing 100080, China; (L.L.); (P.Y.); (Y.D.)
- Correspondence:
| |
Collapse
|
5
|
Informational Masking Effects of Similarity and Uncertainty on Early and Late Stages of Auditory Cortical Processing. Ear Hear 2021; 42:1006-1023. [PMID: 33416259 DOI: 10.1097/aud.0000000000000997] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
PURPOSE Understanding speech in a background of other people talking is a difficult listening situation for hearing-impaired individuals, and even for those with normal hearing. Speech-on-speech masking is known to contribute to increased perceptual difficulty over nonspeech background noise because of informational masking provided over and above the effects of energetic masking. While informational masking research has identified factors of similarity and uncertainty between target and masker that contribute to reduced behavioral performance in speech background noise, critical gaps in knowledge including the underlying neural-perceptual processes remain. By systematically manipulating aspects of acoustic similarity and uncertainty in the same auditory paradigm, the current study examined the time course and objectively quantified these informational masking effects at both early and late stages of auditory processing using auditory evoked potentials (AEPs). METHOD Thirty participants were included in a cross-sectional repeated measures design. Target-masker similarity was manipulated by varying the linguistic/phonetic similarity (i.e., language) of the talkers in the background. Specifically, four levels representing hypothesized increasing levels of informational masking were implemented: (1) no masker (quiet); (2) Mandarin; (3) Dutch; and (4) English. Stimulus uncertainty was manipulated by task complexity, specifically presentation of target-to-target interval (TTI) in the auditory evoked paradigm. Participants had to discriminate between English word stimuli (/bæt/ and /pæt/) presented in an oddball paradigm under each masker condition pressing buttons to either the target or standard stimulus. Responses were recorded simultaneously for P1-N1-P2 (standard waveform) and P3 (target waveform). This design allowed for simultaneous recording of multiple AEP peaks, as well as accuracy, reaction time, and d' behavioral discrimination to button press responses. RESULTS Several trends in AEP components were consistent with effects of increasing linguistic/phonetic similarity and stimulus uncertainty. All babble maskers significantly affected outcomes compared to quiet. In addition, the native language English masker had the largest effect on outcomes in the AEP paradigm, including reduced P3 amplitude and area, as well as decreased accuracy and d' behavioral discrimination to target word responses. AEP outcomes for the Mandarin and Dutch maskers, however, were not significantly different across any measured component. Latency outcomes for both N1 and P3 also supported an effect of stimulus uncertainty, consistent with increased processing time related to greater task complexity. An unanticipated result was the absence of the interaction of linguistic/phonetic similarity and stimulus uncertainty. CONCLUSIONS Observable effects of both similarity and uncertainty were evidenced at a level of the P3 more than the earlier N1 level of auditory cortical processing suggesting that higher-level active auditory processing may be more sensitive to informational masking deficits. The lack of significant interaction between similarity and uncertainty at either level of processing suggests that these informational masking factors operated independently. Speech babble maskers across languages altered AEP component measures, behavioral detection, and reaction time. Specifically, this occurred when the babble was in the native/same language as the target, while the effects of foreign language maskers did not differ. The objective results from this study provide a foundation for further investigation of how the linguistic content of target and masker and task difficulty contribute to difficulty understanding speech-in-noise.
Collapse
|
6
|
Zhang C, Tao R, Zhao H. Auditory spatial attention modulates the unmasking effect of perceptual separation in a "cocktail party" environment. Neuropsychologia 2019; 124:108-116. [PMID: 30659864 DOI: 10.1016/j.neuropsychologia.2019.01.009] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2018] [Revised: 11/01/2018] [Accepted: 01/15/2019] [Indexed: 11/30/2022]
Abstract
The perceptual separation between a signal speech and a competing speech (masker), induced by the precedence effect, plays an important role in releasing the signal speech from the masker, especially in a reverberant environment. The perceptual-separation-induced unmasking effect has been suggested to involve multiple cognitive processes, such as selective attention. However, whether listeners' spatial attention modulate the perceptual-separation-induced unmasking effect is not clear. The present study investigated how perceptual separation and auditory spatial attention interact with each other to facilitate speech perception under a simulated noisy and reverberant environment by analyzing the cortical auditory evoked potentials to the signal speech. The results showed that the N1 wave was significantly enhanced by perceptual separation between the signal and masker regardless of whether the participants' spatial attention was directed to the signal or not. However, the P2 wave was significantly enhanced by perceptual separation only when the participants attended to the signal speech. The results indicate that the perceptual-separation-induced facilitation of P2 needs more attentional resource than that of N1. The results also showed that the signal speech caused an enhanced N1 in the contralateral hemisphere regardless of whether participants' attention was directed to the signal or not. In contrast, the signal speech caused an enhanced P2 in the contralateral hemisphere only when the participant attended to the signal. The results indicate that the hemispheric distribution of N1 is mainly affected by the perceptual features of the acoustic stimuli, while that of P2 is affected by the listeners' attentional status.
Collapse
Affiliation(s)
- Changxin Zhang
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China.
| | - Renxia Tao
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China
| | - Hang Zhao
- Faculty of Education, East China Normal University, Shanghai, China; Key Laboratory of Speech and Hearing Science, East China Normal University, Shanghai, China
| |
Collapse
|
7
|
Kamourieh S, Braga RM, Leech R, Mehta A, Wise RJS. Speech Registration in Symptomatic Memory Impairment. Front Aging Neurosci 2018; 10:201. [PMID: 30038566 PMCID: PMC6046456 DOI: 10.3389/fnagi.2018.00201] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2017] [Accepted: 06/13/2018] [Indexed: 11/20/2022] Open
Abstract
Background: An inability to recall recent conversations often indicates impaired episodic memory retrieval. It may also reflect a failure of attentive registration of spoken sentences which leads to unsuccessful memory encoding. The hypothesis was that patients complaining of impaired memory would demonstrate impaired function of “multiple demand” (MD) brain regions, whose activation profile generalizes across cognitive domains, during speech registration in naturalistic listening conditions. Methods: Using functional MRI, brain activity was measured in 22 normal participants and 31 patients complaining of memory impairment, 21 of whom had possible or probable Alzheimer’s disease (AD). Participants heard a target speaker, either speaking alone or in the presence of distracting background speech, followed by a question to determine if the target speech had been registered. Results: Patients performed poorly at registering verbal information, which correlated with their scores on a screening test of cognitive impairment. Speech registration was associated with widely distributed activity in both auditory cortex and in MD cortex. Additional regions were most active when the target speech had to be separated from background speech. Activity in midline and lateral frontal MD cortex was reduced in the patients. A central cholinesterase inhibitor to increase brain acetylcholine levels in half the patients was not observed to alter brain activity or improve task performance at a second fMRI scan performed 6–11 weeks later. However, individual performances spontaneously fluctuated between the two scanning sessions, and these performance differences correlated with activity within a right hemisphere fronto-temporal system previously associated with sustained auditory attention. Conclusions: Midline and lateralized frontal regions that are engaged in task-dependent attention to, and registration of, verbal information are potential targets for transcranial brain stimulation to improve speech registration in neurodegenerative conditions.
Collapse
Affiliation(s)
- Salwa Kamourieh
- Computational, Cognitive, and Clinical Neuroimaging Laboratory, Division of Brain Sciences, Imperial College London, Hammersmith Hospital, London, United Kingdom
| | - Rodrigo M Braga
- Computational, Cognitive, and Clinical Neuroimaging Laboratory, Division of Brain Sciences, Imperial College London, Hammersmith Hospital, London, United Kingdom.,Center for Brain Science, Harvard University, Cambridge, MA, United States
| | - Robert Leech
- Computational, Cognitive, and Clinical Neuroimaging Laboratory, Division of Brain Sciences, Imperial College London, Hammersmith Hospital, London, United Kingdom
| | - Amrish Mehta
- Department of Neuroradiology, Charing Cross Hospital, Imperial College Healthcare NHS Trust, Faculty of Medicine, Imperial College London, London, United Kingdom
| | - Richard J S Wise
- Computational, Cognitive, and Clinical Neuroimaging Laboratory, Division of Brain Sciences, Imperial College London, Hammersmith Hospital, London, United Kingdom
| |
Collapse
|