1
|
Whitley A, Naylor G, Hadley LV. Used to Be a Dime, Now It's a Dollar: Revised Speech Perception in Noise Key Word Predictability Revisited 40 Years On. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:1229-1242. [PMID: 38563688 PMCID: PMC11005954 DOI: 10.1044/2024_jslhr-23-00615] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Revised: 12/11/2023] [Accepted: 01/12/2024] [Indexed: 04/04/2024]
Abstract
PURPOSE Almost 40 years after its development, in this article, we reexamine the relevance and validity of the ubiquitously used Revised Speech Perception in Noise (R-SPiN) sentence corpus. The R-SPiN corpus includes "high-context" and "low-context" sentences and has been widely used in the field of hearing research to examine the benefit derived from semantic context across English-speaking listeners, but research investigating age differences has yielded somewhat inconsistent findings. We assess the appropriateness of the corpus for use today in different English-language cultures (i.e., British and American) as well as for older and younger adults. METHOD Two hundred forty participants, including older (60-80 years) and younger (19-31 years) adult groups in the the United Kingdom and United States, completed a cloze task consisting of R-SPiN sentences with the final word removed. Cloze, as a measure of predictability, and entropy, as a measure of response uncertainty, were compared between culture and age groups. RESULTS Most critically, of the 200 "high-context" stimuli, only around half were assessed as highly predictable for older adults (United Kingdom: 109; United States: 107); and fewer still, for younger adults (United Kingdom: 75; United States: 81). We also found dominant responses to these "high-context" stimuli varied between cultures, with U.S. responses being more likely to match the original R-SPiN target. CONCLUSIONS Our findings highlight the issue of incomplete transferability of corpus items across English-language cultures as well as diminished equivalency for older and younger adults. By identifying relevant items for each population, this work could facilitate the interpretation of inconsistent findings in the literature, particularly relating to age effects.
Collapse
Affiliation(s)
- Alexina Whitley
- Hearing Sciences – Scottish Section, University of Nottingham, United Kingdom
| | - Graham Naylor
- Hearing Sciences – Scottish Section, University of Nottingham, United Kingdom
| | - Lauren V. Hadley
- Hearing Sciences – Scottish Section, University of Nottingham, United Kingdom
| |
Collapse
|
2
|
Choi HJ, Kyong JS, Won JH, Shim HJ. Effect of spectral degradation on speech intelligibility and cortical representation. Front Neurosci 2024; 18:1368641. [PMID: 38646607 PMCID: PMC11027739 DOI: 10.3389/fnins.2024.1368641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Accepted: 03/25/2024] [Indexed: 04/23/2024] Open
Abstract
Noise-vocoded speech has long been used to investigate how acoustic cues affect speech understanding. Studies indicate that reducing the number of spectral channel bands diminishes speech intelligibility. Despite previous studies examining the channel band effect using earlier event-related potential (ERP) components, such as P1, N1, and P2, a clear consensus or understanding remains elusive. Given our hypothesis that spectral degradation affects higher-order processing of speech understanding beyond mere perception, we aimed to objectively measure differences in higher-order abilities to discriminate or interpret meaning. Using an oddball paradigm with speech stimuli, we examined how neural signals correlate with the evaluation of speech stimuli based on the number of channel bands measuring N2 and P3b components. In 20 young participants with normal hearing, we measured speech intelligibility and N2 and P3b responses using a one-syllable task paradigm with animal and non-animal stimuli across four vocoder conditions with 4, 8, 16, or 32 channel bands. Behavioral data from word repetition clearly affected the number of channel bands, and all pairs were significantly different (p < 0.001). We also observed significant effects of the number of channels on the peak amplitude [F(2.006, 38.117) = 9.077, p < 0.001] and peak latency [F(3, 57) = 26.642, p < 0.001] of the N2 component. Similarly, the P3b component showed significant main effects of the number of channel bands on the peak amplitude [F(2.231, 42.391) = 13.045, p < 0.001] and peak latency [F(3, 57) = 2.968, p = 0.039]. In summary, our findings provide compelling evidence that spectral channel bands profoundly influence cortical speech processing, as reflected in the N2 and P3b components, a higher-order cognitive process. We conclude that spectrally degraded one-syllable speech primarily affects cortical responses during semantic integration.
Collapse
Affiliation(s)
- Hyo Jung Choi
- Department of Otorhinolaryngology-Head and Neck Surgery, Nowon Eulji Medical Center, Eulji University School of Medicine, Seoul, Republic of Korea
- Eulji Tinnitus and Hearing Research Institute, Nowon Eulji Medical Center, Seoul, Republic of Korea
| | - Jeong-Sug Kyong
- Sensory-Organ Research Institute, Medical Research Center, Seoul National University School of Medicine, Seoul, Republic of Korea
- Department of Radiology, Konkuk University Medical Center, Seoul, Republic of Korea
| | - Jong Ho Won
- Hyman, Phelps and McNamara, P.C., Washington, DC, United States
| | - Hyun Joon Shim
- Department of Otorhinolaryngology-Head and Neck Surgery, Nowon Eulji Medical Center, Eulji University School of Medicine, Seoul, Republic of Korea
- Eulji Tinnitus and Hearing Research Institute, Nowon Eulji Medical Center, Seoul, Republic of Korea
| |
Collapse
|
3
|
Ershaid H, Lizarazu M, McLaughlin D, Cooke M, Simantiraki O, Koutsogiannaki M, Lallier M. Contributions of listening effort and intelligibility to cortical tracking of speech in adverse listening conditions. Cortex 2024; 172:54-71. [PMID: 38215511 DOI: 10.1016/j.cortex.2023.11.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Revised: 09/05/2023] [Accepted: 11/14/2023] [Indexed: 01/14/2024]
Abstract
Cortical tracking of speech is vital for speech segmentation and is linked to speech intelligibility. However, there is no clear consensus as to whether reduced intelligibility leads to a decrease or an increase in cortical speech tracking, warranting further investigation of the factors influencing this relationship. One such factor is listening effort, defined as the cognitive resources necessary for speech comprehension, and reported to have a strong negative correlation with speech intelligibility. Yet, no studies have examined the relationship between speech intelligibility, listening effort, and cortical tracking of speech. The aim of the present study was thus to examine these factors in quiet and distinct adverse listening conditions. Forty-nine normal hearing adults listened to sentences produced casually, presented in quiet and two adverse listening conditions: cafeteria noise and reverberant speech. Electrophysiological responses were registered with electroencephalogram, and listening effort was estimated subjectively using self-reported scores and objectively using pupillometry. Results indicated varying impacts of adverse conditions on intelligibility, listening effort, and cortical tracking of speech, depending on the preservation of the speech temporal envelope. The more distorted envelope in the reverberant condition led to higher listening effort, as reflected in higher subjective scores, increased pupil diameter, and stronger cortical tracking of speech in the delta band. These findings suggest that using measures of listening effort in addition to those of intelligibility is useful for interpreting cortical tracking of speech results. Moreover, reading and phonological skills of participants were positively correlated with listening effort in the cafeteria condition, suggesting a special role of expert language skills in processing speech in this noisy condition. Implications for future research and theories linking atypical cortical tracking of speech and reading disorders are further discussed.
Collapse
Affiliation(s)
- Hadeel Ershaid
- Basque Center on Cognition, Brain and Language, San Sebastian, Spain.
| | - Mikel Lizarazu
- Basque Center on Cognition, Brain and Language, San Sebastian, Spain.
| | - Drew McLaughlin
- Basque Center on Cognition, Brain and Language, San Sebastian, Spain.
| | - Martin Cooke
- Ikerbasque, Basque Science Foundation, Bilbao, Spain.
| | | | | | - Marie Lallier
- Basque Center on Cognition, Brain and Language, San Sebastian, Spain; Ikerbasque, Basque Science Foundation, Bilbao, Spain.
| |
Collapse
|
4
|
Jiang J, Johnson JCS, Requena-Komuro MC, Benhamou E, Sivasathiaseelan H, Chokesuwattanaskul A, Nelson A, Nortley R, Weil RS, Volkmer A, Marshall CR, Bamiou DE, Warren JD, Hardy CJD. Comprehension of acoustically degraded speech in Alzheimer's disease and primary progressive aphasia. Brain 2023; 146:4065-4076. [PMID: 37184986 PMCID: PMC10545509 DOI: 10.1093/brain/awad163] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 04/20/2023] [Accepted: 04/27/2023] [Indexed: 05/17/2023] Open
Abstract
Successful communication in daily life depends on accurate decoding of speech signals that are acoustically degraded by challenging listening conditions. This process presents the brain with a demanding computational task that is vulnerable to neurodegenerative pathologies. However, despite recent intense interest in the link between hearing impairment and dementia, comprehension of acoustically degraded speech in these diseases has been little studied. Here we addressed this issue in a cohort of 19 patients with typical Alzheimer's disease and 30 patients representing the three canonical syndromes of primary progressive aphasia (non-fluent/agrammatic variant primary progressive aphasia; semantic variant primary progressive aphasia; logopenic variant primary progressive aphasia), compared to 25 healthy age-matched controls. As a paradigm for the acoustically degraded speech signals of daily life, we used noise-vocoding: synthetic division of the speech signal into frequency channels constituted from amplitude-modulated white noise, such that fewer channels convey less spectrotemporal detail thereby reducing intelligibility. We investigated the impact of noise-vocoding on recognition of spoken three-digit numbers and used psychometric modelling to ascertain the threshold number of noise-vocoding channels required for 50% intelligibility by each participant. Associations of noise-vocoded speech intelligibility threshold with general demographic, clinical and neuropsychological characteristics and regional grey matter volume (defined by voxel-based morphometry of patients' brain images) were also assessed. Mean noise-vocoded speech intelligibility threshold was significantly higher in all patient groups than healthy controls, and significantly higher in Alzheimer's disease and logopenic variant primary progressive aphasia than semantic variant primary progressive aphasia (all P < 0.05). In a receiver operating characteristic analysis, vocoded intelligibility threshold discriminated Alzheimer's disease, non-fluent variant and logopenic variant primary progressive aphasia patients very well from healthy controls. Further, this central hearing measure correlated with overall disease severity but not with peripheral hearing or clear speech perception. Neuroanatomically, after correcting for multiple voxel-wise comparisons in predefined regions of interest, impaired noise-vocoded speech comprehension across syndromes was significantly associated (P < 0.05) with atrophy of left planum temporale, angular gyrus and anterior cingulate gyrus: a cortical network that has previously been widely implicated in processing degraded speech signals. Our findings suggest that the comprehension of acoustically altered speech captures an auditory brain process relevant to daily hearing and communication in major dementia syndromes, with novel diagnostic and therapeutic implications.
Collapse
Affiliation(s)
- Jessica Jiang
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Jeremy C S Johnson
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Maï-Carmen Requena-Komuro
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
- Kidney Cancer Program, UT Southwestern Medical Centre, Dallas, TX 75390, USA
| | - Elia Benhamou
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Harri Sivasathiaseelan
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Anthipa Chokesuwattanaskul
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
- Division of Neurology, Department of Internal Medicine, King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok 10330, Thailand
| | - Annabel Nelson
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Ross Nortley
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
- Wexham Park Hospital, Frimley Health NHS Foundation Trust, Slough SL2 4HL, UK
| | - Rimona S Weil
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Anna Volkmer
- Division of Psychology and Language Sciences, University College London, London WC1H 0AP, UK
| | - Charles R Marshall
- Preventive Neurology Unit, Wolfson Institute of Population Health, Queen Mary University of London, London EC1M 6BQ, UK
| | - Doris-Eva Bamiou
- UCL Ear Institute and UCL/UCLH Biomedical Research Centre, National Institute of Health Research, University College London, London WC1X 8EE, UK
| | - Jason D Warren
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| | - Chris J D Hardy
- Dementia Research Centre, Department of Neurodegenerative Disease, UCL Queen Square Institute of Neurology, University College London, London WC1N 3AR, UK
| |
Collapse
|
5
|
Song J, Kim B, Kim M, Iverson P. The Korean Speech Recognition Sentences: A Large Corpus for Evaluating Semantic Context and Language Experience in Speech Perception. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:3399-3412. [PMID: 37672785 PMCID: PMC10558151 DOI: 10.1044/2023_jslhr-23-00137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/24/2023] [Revised: 04/29/2023] [Accepted: 05/25/2023] [Indexed: 09/08/2023]
Abstract
PURPOSE The aim of this study was to develop and validate a large Korean sentence set with varying degrees of semantic predictability that can be used for testing speech recognition and lexical processing. METHOD Sentences differing in the degree of final-word predictability (predictable, neutral, and anomalous) were created with words selected to be suitable for both native and nonnative speakers of Korean. Semantic predictability was evaluated through a series of cloze tests in which native (n = 56) and nonnative (n = 19) speakers of Korean participated. This study also used a computer language model to evaluate final-word predictabilities; this is a novel approach that the current study adopted to reduce human effort in validating a large number of sentences, which produced results comparable to those of the cloze tests. In a speech recognition task, the sentences were presented to native (n = 23) and nonnative (n = 21) speakers of Korean in speech-shaped noise at two levels of noise. RESULTS The results of the speech-in-noise experiment demonstrated that the intelligibility of the sentences was similar to that of related English corpora. That is, intelligibility was significantly different depending on the semantic condition, and the sentences had the right degree of difficulty for assessing intelligibility differences depending on noise levels and language experience. CONCLUSIONS This corpus (1,021 sentences in total) adds to the target languages available in speech research and will allow researchers to investigate a range of issues in speech perception in Korean. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.24045582.
Collapse
Affiliation(s)
- Jieun Song
- School of Digital Humanities and Computational Social Sciences, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Byungjun Kim
- Center for Digital Humanities and Computational Social Sciences, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Minjeong Kim
- Graduate School of Culture Technology, Korea Advanced Institute of Science and Technology, Daejeon, South Korea
| | - Paul Iverson
- Department of Speech, Hearing and Phonetic Sciences, University College London, United Kingdom
| |
Collapse
|
6
|
Kovács P, Szalárdy O, Winkler I, Tóth B. Two effects of perceived speaker similarity in resolving the cocktail party situation - ERPs and functional connectivity. Biol Psychol 2023; 182:108651. [PMID: 37517603 DOI: 10.1016/j.biopsycho.2023.108651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Revised: 07/15/2023] [Accepted: 07/24/2023] [Indexed: 08/01/2023]
Abstract
Following a speaker in multi-talker environments requires the listener to separate the speakers' voices and continuously focus attention on one speech stream. While the dissimilarity of voices may make speaker separation easier, it may also affect maintaining the focus of attention. To assess these effects, electrophysiological (EEG) and behavioral data were collected from healthy young adults while they listened to two concurrent speech streams performing an online lexical detection task and an offline recognition memory task. Perceptual speaker similarity was manipulated on four levels: identical, similar, dissimilar, and opposite-gender speakers. Behavioral and electrophysiological data suggested that, while speaker similarity hinders auditory stream segregation, dissimilarity hinders maintaining the focus of attention by making the to-be-ignored speech stream more distracting. Thus, resolving the cocktail party situation poses different problems at different levels of perceived speaker similarity, resulting in different listening strategies.
Collapse
Affiliation(s)
- Petra Kovács
- Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Budapest, Hungary; Department of Cognitive Science, Budapest University of Technology and Economics, Budapest, Hungary
| | - Orsolya Szalárdy
- Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Budapest, Hungary; Institute of Behavioural Sciences, Faculty of Medicine, Semmelweis University, Budapest, Hungary
| | - István Winkler
- Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Budapest, Hungary
| | - Brigitta Tóth
- Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Budapest, Hungary.
| |
Collapse
|
7
|
Wisniewski MG, Zakrzewski AC. Effortful listening produces both enhancement and suppression of alpha in the EEG. AUDITORY PERCEPTION & COGNITION 2023; 6:289-299. [PMID: 38665905 PMCID: PMC11044958 DOI: 10.1080/25742442.2023.2218239] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/20/2023] [Accepted: 05/18/2023] [Indexed: 04/28/2024]
Abstract
Introduction Adverse listening conditions can drive increased mental effort during listening. Neuromagnetic alpha oscillations (8-13 Hz) may index this listening effort, but inconsistencies regarding the direction of the relationship are abundant. We performed source analyses on high-density EEG data collected during a speech-on-speech listening task to address the possibility that opposing alpha power relationships among alpha producing brain sources drive this inconsistency. Methods Listeners (N=20) heard two simultaneously presented sentences of the form: Ready go to now. They either reported the color/number pair of a "Baron" call sign sentence (active: high effort), or ignored the stimuli (passive: low effort). Independent component analysis (ICA) was used to segregate temporally distinct sources in the EEG. Results Analysis of independent components (ICs) revealed simultaneous alpha enhancements (e.g., for somatomotor mu ICs) and suppressions (e.g., for left temporal ICs) for different brain sources. The active condition exhibited stronger enhancement for left somatomotor mu rhythm ICs, but stronger suppression for central occipital ICs. Discussion This study shows both alpha enhancement and suppression to be associated with increases in listening effort. Literature inconsistencies could partially relate to some source activities overwhelming others in scalp recordings.
Collapse
Affiliation(s)
- Matthew G. Wisniewski
- Department of Psychological Sciences, Kansas State University, Manhattan, Kansas, USA
| | | |
Collapse
|
8
|
Sen O, Sheehan AM, Raman PR, Khara KS, Khalifa A, Chatterjee B. Machine-Learning Methods for Speech and Handwriting Detection Using Neural Signals: A Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:5575. [PMID: 37420741 DOI: 10.3390/s23125575] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Revised: 06/09/2023] [Accepted: 06/12/2023] [Indexed: 07/09/2023]
Abstract
Brain-Computer Interfaces (BCIs) have become increasingly popular in recent years due to their potential applications in diverse fields, ranging from the medical sector (people with motor and/or communication disabilities), cognitive training, gaming, and Augmented Reality/Virtual Reality (AR/VR), among other areas. BCI which can decode and recognize neural signals involved in speech and handwriting has the potential to greatly assist individuals with severe motor impairments in their communication and interaction needs. Innovative and cutting-edge advancements in this field have the potential to develop a highly accessible and interactive communication platform for these people. The purpose of this review paper is to analyze the existing research on handwriting and speech recognition from neural signals. So that the new researchers who are interested in this field can gain thorough knowledge in this research area. The current research on neural signal-based recognition of handwriting and speech has been categorized into two main types: invasive and non-invasive studies. We have examined the latest papers on converting speech-activity-based neural signals and handwriting-activity-based neural signals into text data. The methods of extracting data from the brain have also been discussed in this review. Additionally, this review includes a brief summary of the datasets, preprocessing techniques, and methods used in these studies, which were published between 2014 and 2022. This review aims to provide a comprehensive summary of the methodologies used in the current literature on neural signal-based recognition of handwriting and speech. In essence, this article is intended to serve as a valuable resource for future researchers who wish to investigate neural signal-based machine-learning methods in their work.
Collapse
Affiliation(s)
- Ovishake Sen
- Department of ECE, University of Florida, Gainesville, FL 32611, USA
| | - Anna M Sheehan
- Department of ECE, University of Florida, Gainesville, FL 32611, USA
| | - Pranay R Raman
- Department of ECE, University of Florida, Gainesville, FL 32611, USA
| | - Kabir S Khara
- Department of ECE, University of Florida, Gainesville, FL 32611, USA
| | - Adam Khalifa
- Department of ECE, University of Florida, Gainesville, FL 32611, USA
| | | |
Collapse
|
9
|
Stinkeste C, Vincent MA, Delrue L, Brunellière A. Between alpha and gamma oscillations: Neural signatures of linguistic predictions and listener's attention to speaker's communication intention. Biol Psychol 2023; 180:108583. [PMID: 37156325 DOI: 10.1016/j.biopsycho.2023.108583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2022] [Revised: 05/04/2023] [Accepted: 05/05/2023] [Indexed: 05/10/2023]
Abstract
When listeners hear a message produced by their interlocutor, they can predict upcoming words thanks to the sentential context and their attention can be focused on the speaker's communication intention. In two electroencephalographical (EEG) studies, we investigated the oscillatory correlates of prediction in spoken-language comprehension and how they are modulated by the listener's attention. Sentential contexts which were strongly predictive of a particular word were ended by a possessive adjective either matching the gender of the predicted word or not. Alpha, beta and gamma oscillations were studied as they were considered to play a crucial role in the predictive process. While evidence of word prediction was related to alpha fluctuations when listeners focused their attention on sentence meaning, changes in high-gamma oscillations were triggered by word prediction when listeners focused their attention on the speaker's communication intention. Independently of the endogenous attention to a level of linguistic information, the oscillatory correlates of word predictions in language comprehension were sensitive to the prosodic emphasis produced by the speaker at a late stage. These findings thus bear major implications for understanding the neural mechanisms that support predictive processing in spoken-language comprehension.
Collapse
Affiliation(s)
- Charlotte Stinkeste
- Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France
| | - Marion A Vincent
- Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France
| | - Laurence Delrue
- Univ. Lille, CNRS, UMR 8163 - STL - Savoirs Textes Langage, F-59000 Lille, France
| | - Angèle Brunellière
- Univ. Lille, CNRS, UMR 9193 - SCALab - Sciences Cognitives et Sciences Affectives, F-59000 Lille, France.
| |
Collapse
|
10
|
Kaufman M, Zion Golumbic E. Listening to two speakers: Capacity and tradeoffs in neural speech tracking during Selective and Distributed Attention. Neuroimage 2023; 270:119984. [PMID: 36854352 DOI: 10.1016/j.neuroimage.2023.119984] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Revised: 02/06/2023] [Accepted: 02/24/2023] [Indexed: 02/27/2023] Open
Abstract
Speech comprehension is severely compromised when several people talk at once, due to limited perceptual and cognitive resources. In such circumstances, top-down attention mechanisms can actively prioritize processing of task-relevant speech. However, behavioral and neural evidence suggest that this selection is not exclusive, and the system may have sufficient capacity to process additional speech input as well. Here we used a data-driven approach to contrast two opposing hypotheses regarding the system's capacity to co-represent competing speech: Can the brain represent two speakers equally or is the system fundamentally limited, resulting in tradeoffs between them? Neural activity was measured using magnetoencephalography (MEG) as human participants heard concurrent speech narratives and engaged in two tasks: Selective Attention, where only one speaker was task-relevant and Distributed Attention, where both speakers were equally relevant. Analysis of neural speech-tracking revealed that both tasks engaged a similar network of brain regions involved in auditory processing, attentional control and speech processing. Interestingly, during both Selective and Distributed Attention the neural representation of competing speech showed a bias towards one speaker. This is in line with proposed 'bottlenecks' for co-representation of concurrent speech and suggests that good performance on distributed attention tasks may be achieved by toggling attention between speakers over time.
Collapse
Affiliation(s)
- Maya Kaufman
- The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, Israel
| | - Elana Zion Golumbic
- The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, Israel.
| |
Collapse
|
11
|
Kovács P, Tóth B, Honbolygó F, Szalárdy O, Kohári A, Mády K, Magyari L, Winkler I. Speech prosody supports speaker selection and auditory stream segregation in a multi-talker situation. Brain Res 2023; 1805:148246. [PMID: 36657631 DOI: 10.1016/j.brainres.2023.148246] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Revised: 01/06/2023] [Accepted: 01/12/2023] [Indexed: 01/19/2023]
Abstract
To process speech in a multi-talker environment, listeners need to segregate the mixture of incoming speech streams and focus their attention on one of them. Potentially, speech prosody could aid the segregation of different speakers, the selection of the desired speech stream, and detecting targets within the attended stream. For testing these issues, we recorded behavioral responses and extracted event-related potentials and functional brain networks from electroencephalographic signals recorded while participants listened to two concurrent speech streams, performing a lexical detection and a recognition memory task in parallel. Prosody manipulation was applied to the attended speech stream in one group of participants and to the ignored speech stream in another group. Naturally recorded speech stimuli were either intact, synthetically F0-flattened, or prosodically suppressed by the speaker. Results show that prosody - especially the parsing cues mediated by speech rate - facilitates stream selection, while playing a smaller role in auditory stream segmentation and target detection.
Collapse
Affiliation(s)
- Petra Kovács
- Department of Cognitive Science, Budapest University of Technology and Economics, Hungary
| | - Brigitta Tóth
- Institute of Cognitive Neuroscience and Psychology, Research Center for Natural Sciences, Hungary.
| | - Ferenc Honbolygó
- Brain Imaging Center, Research Center for Natural Sciences, Hungary
| | - Orsolya Szalárdy
- Institute of Cognitive Neuroscience and Psychology, Research Center for Natural Sciences, Hungary; Institute of Behavioural Sciences, Faculty of Medicine, Semmelweis University, Budapest, Hungary
| | - Anna Kohári
- Research Group of Phonetics, Institute for General and Hungarian Linguistics, Hungarian Research Centre for Linguistics, Hungary
| | - Katalin Mády
- Research Group of Phonetics, Institute for General and Hungarian Linguistics, Hungarian Research Centre for Linguistics, Hungary
| | - Lilla Magyari
- Department of Social Studies, Faculty of Social Sciences, University of Stavanger, Stavanger, Norway; Norwegian Centre for Reading Education and Research, Faculty of Arts and Education, University of Stavanger, Stavanger, Norway
| | - István Winkler
- Institute of Cognitive Neuroscience and Psychology, Research Center for Natural Sciences, Hungary
| |
Collapse
|
12
|
Voola M, Wedekind A, Nguyen AT, Marinovic W, Rajan G, Tavora-Vieira D. Event-Related Potentials of Single-Sided Deaf Cochlear Implant Users: Using a Semantic Oddball Paradigm in Noise. Audiol Neurootol 2023; 28:280-293. [PMID: 36940674 PMCID: PMC10413801 DOI: 10.1159/000529485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Accepted: 01/31/2023] [Indexed: 03/23/2023] Open
Abstract
INTRODUCTION In individuals with single-sided deafness (SSD), who are characterised by profound hearing loss in one ear and normal hearing in the contralateral ear, binaural input is no longer present. A cochlear implant (CI) can restore functional hearing in the profoundly deaf ear, with previous literature demonstrating improvements in speech-in-noise intelligibility with the CI. However, we currently have limited understanding of the neural processes involved (e.g., how the brain integrates the electrical signal produced by the CI with the acoustic signal produced by the normal hearing ear) and how modulation of these processes with a CI contributes to improved speech-in-noise intelligibility. Using a semantic oddball paradigm presented in the presence of background noise, this study aims to investigate how the provision of CI impacts speech-in-noise perception of SSD-CI users. METHOD Task performance (reaction time, reaction time variability, target accuracy, subjective listening effort) and high density electroencephalography from twelve SSD-CI participants were recorded, while they completed a semantic acoustic oddball task. Reaction time was defined as the time taken for a participant to press the response button after stimulus onset. All participants completed the oddball task in three different free-field conditions with the speech and noise coming from different speakers. The three tasks were: (1) CI-On in background noise, (2) CI-Off in background noise, and (3) CI-On without background noise (Control). Task performance and electroencephalography data (N2N4 and P3b) were recorded for each condition. Speech in noise and sound localisation ability were also measured. RESULTS Reaction time was significantly different between all tasks with CI-On (M [SE] = 809 [39.9] ms) having faster RTs than CI-Off (M [SE] = 845 [39.9] ms) and Control (M [SE] = 785 [39.9] ms) being the fastest condition. The Control condition exhibited significantly shorter N2N4 and P3b area latency compared to the other two conditions. However, despite these differences noticed in RTs and area latency, we observed similar results between all three conditions for N2N4 and P3b difference area. CONCLUSION The inconsistency between the behavioural and neural results suggests that EEG may not be a reliable measure of cognitive effort. This rationale is further supported by different explanations used in past studies to explain N2N4 and P3b effects. Future studies should look to alternative measures of auditory processing (e.g., pupillometry) to gain a deeper understanding of the underlying auditory processes that facilitate speech-in-noise intelligibility.
Collapse
Affiliation(s)
- Marcus Voola
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA, Australia
- Department of Audiology, Fiona Stanley Fremantle Hospitals Group, Perth, WA, Australia
| | - Andre Wedekind
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA, Australia
- Department of Audiology, Fiona Stanley Fremantle Hospitals Group, Perth, WA, Australia
| | - An T. Nguyen
- School of Population Health, Curtin University, Perth, WA, Australia
| | - Welber Marinovic
- School of Population Health, Curtin University, Perth, WA, Australia
| | - Gunesh Rajan
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA, Australia
- Deptartment of Otolaryngology, Head and Neck Surgery, Luzerner Kantonsspital, Luzern, Switzerland
| | - Dayse Tavora-Vieira
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA, Australia
- Department of Audiology, Fiona Stanley Fremantle Hospitals Group, Perth, WA, Australia
- School of Population Health, Curtin University, Perth, WA, Australia
| |
Collapse
|
13
|
Han JH, Lee J, Lee HJ. The effect of noise on the cortical activity patterns of speech processing in adults with single-sided deafness. Front Neurol 2023; 14:1054105. [PMID: 37006498 PMCID: PMC10060629 DOI: 10.3389/fneur.2023.1054105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2022] [Accepted: 02/27/2023] [Indexed: 03/18/2023] Open
Abstract
The most common complaint in people with single-sided deafness (SSD) is difficulty in understanding speech in a noisy environment. Moreover, the neural mechanism of speech-in-noise (SiN) perception in SSD individuals is still poorly understood. In this study, we measured the cortical activity in SSD participants during a SiN task to compare with a speech-in-quiet (SiQ) task. Dipole source analysis revealed left hemispheric dominance in both left- and right-sided SSD group. Contrary to SiN listening, this hemispheric difference was not found during SiQ listening in either group. In addition, cortical activation in the right-sided SSD individuals was independent of the location of sound whereas activation sites in the left-sided SSD group were altered by the sound location. Examining the neural-behavioral relationship revealed that N1 activation is associated with the duration of deafness and the SiN perception ability of individuals with SSD. Our findings indicate that SiN listening is processed differently in the brains of left and right SSD individuals.
Collapse
Affiliation(s)
- Ji-Hye Han
- Laboratory of Brain and Cognitive Sciences for Convergence Medicine, Hallym University College of Medicine, Anyang, Republic of Korea
- Ear and Interaction Center, Doheun Institute for Digital Innovation in Medicine (D.I.D.I.M.), Hallym University Medical Center, Anyang, Republic of Korea
| | - Jihyun Lee
- Laboratory of Brain and Cognitive Sciences for Convergence Medicine, Hallym University College of Medicine, Anyang, Republic of Korea
- Ear and Interaction Center, Doheun Institute for Digital Innovation in Medicine (D.I.D.I.M.), Hallym University Medical Center, Anyang, Republic of Korea
| | - Hyo-Jeong Lee
- Laboratory of Brain and Cognitive Sciences for Convergence Medicine, Hallym University College of Medicine, Anyang, Republic of Korea
- Ear and Interaction Center, Doheun Institute for Digital Innovation in Medicine (D.I.D.I.M.), Hallym University Medical Center, Anyang, Republic of Korea
- Department of Otorhinolaryngology-Head and Neck Surgery, Hallym University College of Medicine, Chuncheon, Republic of Korea
- *Correspondence: Hyo-Jeong Lee
| |
Collapse
|
14
|
Lubinus C, Keitel A, Obleser J, Poeppel D, Rimmele JM. Explaining flexible continuous speech comprehension from individual motor rhythms. Proc Biol Sci 2023; 290:20222410. [PMID: 36855868 PMCID: PMC9975658 DOI: 10.1098/rspb.2022.2410] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/02/2023] Open
Abstract
When speech is too fast, the tracking of the acoustic signal along the auditory pathway deteriorates, leading to suboptimal speech segmentation and decoding of speech information. Thus, speech comprehension is limited by the temporal constraints of the auditory system. Here we ask whether individual differences in auditory-motor coupling strength in part shape these temporal constraints. In two behavioural experiments, we characterize individual differences in the comprehension of naturalistic speech as function of the individual synchronization between the auditory and motor systems and the preferred frequencies of the systems. Obviously, speech comprehension declined at higher speech rates. Importantly, however, both higher auditory-motor synchronization and higher spontaneous speech motor production rates were predictive of better speech-comprehension performance. Furthermore, performance increased with higher working memory capacity (digit span) and higher linguistic, model-based sentence predictability-particularly so at higher speech rates and for individuals with high auditory-motor synchronization. The data provide evidence for a model of speech comprehension in which individual flexibility of not only the motor system but also auditory-motor synchronization may play a modulatory role.
Collapse
Affiliation(s)
- Christina Lubinus
- Department of Neuroscience and Department of Cognitive Neuropsychology, Max-Planck-Institute for Empirical Aesthetics, 60322 Frankfurt am Main, Germany
| | - Anne Keitel
- Psychology, University of Dundee, Dundee DD1 4HN, UK
| | - Jonas Obleser
- Department of Psychology, University of Lübeck, Lübeck, Germany
- Center for Brain, Behavior, and Metabolism, University of Lübeck, Lübeck, Germany
| | - David Poeppel
- Department of Psychology, New York University, New York, NY, USA
- Max Planck NYU Center for Language, Music, and Emotion, New York, NY, USA
- Ernst Strüngmann Institute for Neuroscience (in Cooperation with Max Planck Society), Frankfurt am Main, Germany
| | - Johanna M. Rimmele
- Department of Neuroscience and Department of Cognitive Neuropsychology, Max-Planck-Institute for Empirical Aesthetics, 60322 Frankfurt am Main, Germany
- Max Planck NYU Center for Language, Music, and Emotion, New York, NY, USA
| |
Collapse
|
15
|
Guediche S, Navarra-Barindelli E, Martin CD. Noise Modulates Crosslinguistic Effects on Second-Language Auditory Word Recognition. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:635-647. [PMID: 36729589 DOI: 10.1044/2022_jslhr-22-00368] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
PURPOSE This study investigates whether crosslinguistic effects on auditory word recognition are modulated by the quality of the auditory signal (clear and noisy). METHOD In an online experiment, a group of Spanish-English bilingual listeners performed an auditory lexical decision task, in their second language, English. Words and pseudowords were either presented in the clear or were embedded in white auditory noise. Target words were varied in the degree to which they overlapped in their phonological form with their translation equivalents and were categorized according to their overlap as cognates (form and meaning) or noncognates (meaning only). In order to test for effects of crosslinguistic competition, the phonological neighborhood density of the targets' translations was also manipulated. RESULTS The results show that crosslinguistic effects are impacted by noise; when the translation had a high neighborhood density, performance was worse for cognates than for noncognates, especially in noise. CONCLUSIONS The findings suggest that noise increases lexical competition across languages, as it does within a language, and that the crosslinguistic phonological overlap for cognates compared with noncognates can further increase the pool of competitors by co-activating crosslinguistic lexical candidates. The results are discussed within the context of the bilingual word recognition literature and models of language and bilingual lexical processing.
Collapse
Affiliation(s)
- Sara Guediche
- BCBL. Basque Center on Cognition, Brain and Language, Donostia-San Sebastian, Gipuzkoa, Spain
| | - Eugenia Navarra-Barindelli
- BCBL. Basque Center on Cognition, Brain and Language, Donostia-San Sebastian, Gipuzkoa, Spain
- Universidad del País Vasco, Donostia-San Sebastian, Gipuzkoa, Spain
| | - Clara D Martin
- BCBL. Basque Center on Cognition, Brain and Language, Donostia-San Sebastian, Gipuzkoa, Spain
- Ikerbasque, Donostia-San Sebastian, Gipuzkoa, Spain
| |
Collapse
|
16
|
Wang Q, Zhao S, He Z, Zhang S, Jiang X, Zhang T, Liu T, Liu C, Han J. Modeling functional difference between gyri and sulci within intrinsic connectivity networks. Cereb Cortex 2023; 33:933-947. [PMID: 35332916 DOI: 10.1093/cercor/bhac111] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2021] [Revised: 02/17/2022] [Accepted: 02/18/2022] [Indexed: 11/12/2022] Open
Abstract
Recently, the functional roles of the human cortical folding patterns have attracted increasing interest in the neuroimaging community. However, most existing studies have focused on the gyro-sulcal functional relationship on a whole-brain scale but possibly overlooked the localized and subtle functional differences of brain networks. Actually, accumulating evidences suggest that functional brain networks are the basic unit to realize the brain function; thus, the functional relationships between gyri and sulci still need to be further explored within different functional brain networks. Inspired by these evidences, we proposed a novel intrinsic connectivity network (ICN)-guided pooling-trimmed convolutional neural network (I-ptFCN) to revisit the functional difference between gyri and sulci. By testing the proposed model on the task functional magnetic resonance imaging (fMRI) datasets of the Human Connectome Project, we found that the classification accuracy of gyral and sulcal fMRI signals varied significantly for different ICNs, indicating functional heterogeneity of cortical folding patterns in different brain networks. The heterogeneity may be contributed by sulci, as only sulcal signals show heterogeneous frequency features across different ICNs, whereas the frequency features of gyri are homogeneous. These results offer novel insights into the functional difference between gyri and sulci and enlighten the functional roles of cortical folding patterns.
Collapse
Affiliation(s)
- Qiyu Wang
- School of Automation, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| | - Shijie Zhao
- School of Automation, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| | - Zhibin He
- School of Automation, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| | - Shu Zhang
- School of Computer Science, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| | - Xi Jiang
- School of Life Science and Technology, MOE Key Lab for Neuroinformation, University of Electronic Science and Technology of China, Chengdu, Sichuan 611731, China
| | - Tuo Zhang
- School of Automation, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| | - Tianming Liu
- Cortical Architecture Imaging and Discovery Lab, Department of Computer Science and Bioimaging Research Center, The University of Georgia, Athens, GA 30605, United States
| | - Cirong Liu
- CAS Center for Excellence in Brain Science and Intelligence Technology, Institute of Neuroscience, Chinese Academy of Sciences, Shanghai 200031, China
| | - Junwei Han
- School of Automation, Northwestern Polytechnical University, Xi'an, Shaanxi 710072, China
| |
Collapse
|
17
|
A Study of Event-Related Potentials During Monaural and Bilateral Hearing in Single-Sided Deaf Cochlear Implant Users. Ear Hear 2023:00003446-990000000-00102. [PMID: 36706105 DOI: 10.1097/aud.0000000000001326] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
OBJECTIVES Single-sided deafness (SSD) is characterized by a profoundly deaf ear and normal hearing in the contralateral ear. A cochlear implant (CI) is the only method to restore functional hearing in a profoundly deaf ear. In a previous study, we identified that the cortical processing of a CI signal differs from the normal-hearing ear (NHE) when directly compared using an auditory oddball paradigm consisting of pure tones. However, exactly how the brain integrates the electrical and acoustic signal is not well investigated. This study aims to understand how the provision of the CI in combination with the NHE may improve SSD CI users' ability to discriminate and evaluate auditory stimuli. DESIGN Electroencephalography from 10 SSD-CI participants (4 participated in the previous pure-tone study) were recorded during a semantic acoustic oddball task, where they were required to discriminate between odd and even numbers. Stimuli were presented in four hearing conditions: directly through the CI, directly to the NHE, or in free field with the CI switched on and off. We examined task-performance (response time and accuracy) and measured N1, P2, N2N4, and P3b event-related brain potentials (ERPs) linked to the detection, discrimination, and evaluation of task relevant stimuli. Sound localization and speech in noise comprehension was also examined. RESULTS In direct presentation, task performance was superior during NHE compared with CI (shorter and less varied reaction times [~720 versus ~842 msec], higher target accuracy [~93 versus ~70%]) and early neural responses (N1 and P2) were enhanced for NHE suggesting greater signal saliency. However, the size of N2N4 and P3b target-standard effects did not differ significantly between NHE and CI. In free field, target accuracy was similarly high with the CI (FF-On) and without the CI (FF-Off) (~95%), with some evidence of CI interference during FF-On (more variable and slightly but significantly delayed reaction times [~737 versus ~709 msec]). Early neural responses and late effects were also greater during FF-On. Performance on sound localization and speech in noise comprehension (S CI N NHE configuration only) was significantly greater during FF-On. CONCLUSIONS Both behavioral and neural responses in the semantic oddball task were sensitive to CI in both direct and free-field presentations. Direct conditions revealed that participants could perform the task with the CI alone, although performance was suboptimal and early neural responses were reduced when compared with the NHE. For free-field, the addition of the CI was associated with enhanced early and late neural responses, but this did not result in improved task performance. Enhanced neural responses show that the additional input from the CI is modulating relevant perceptual and cognitive processes, but the benefit of binaural hearing on behavior may not be realized in simple oddball tasks which can be adequately performed with the NHE. Future studies interested in binaural hearing should examine performance under noisy conditions and/or use spatial cues to allow headroom for the measurement of binaural benefit.
Collapse
|
18
|
Gohari N, Dastgerdi ZH, Rouhbakhsh N, Afshar S, Mobini R. Training Programs for Improving Speech Perception in Noise: A Review. J Audiol Otol 2023; 27:1-9. [PMID: 36710414 PMCID: PMC9884994 DOI: 10.7874/jao.2022.00283] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2022] [Accepted: 10/26/2022] [Indexed: 01/20/2023] Open
Abstract
Understanding speech in the presence of noise is difficult and challenging, even for people with normal hearing. Accurate pitch perception, coding and decoding of temporal and intensity cues, and cognitive factors are involved in speech perception in noise (SPIN); disruption in any of these can be a barrier to SPIN. Because the physiological representations of sounds can be corrected by exercises, training methods for any impairment can be used to improve speech perception. This study describes the various types of bottom-up training methods: pitch training based on fundamental frequency (F0) and harmonics; spatial, temporal, and phoneme training; and top-down training methods, such as cognitive training of functional memory. This study also discusses music training that affects both bottom-up and top-down components and speech training in noise. Given the effectiveness of all these training methods, we recommend identifying the defects underlying SPIN disorders and selecting the best training approach.
Collapse
Affiliation(s)
- Nasrin Gohari
- Hearing Disorders Research Center, Department of Audiology, School of Rehabilitation, Hamadan University of Medical Sciences, Hamadan, Iran
| | - Zahra Hosseini Dastgerdi
- Department of Audiology, School of Rehabilitation, Isfahan University of Medical Sciences, Isfahan, Iran,Address for correspondence Zahra Hosseini Dastgerdi, PhD Department of Audiology, School of Rehabilitation, Isfahan University of Medical Sciences, Isfahan, Iran Tel +98-09132947800 Fax +98-(311)5145-668 E-mail
| | - Nematollah Rouhbakhsh
- Department of Audiology, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| | - Sara Afshar
- Hearing Disorders Research Center, Department of Audiology, School of Rehabilitation, Hamadan University of Medical Sciences, Hamadan, Iran
| | - Razieh Mobini
- Hearing Disorders Research Center, Department of Audiology, School of Rehabilitation, Hamadan University of Medical Sciences, Hamadan, Iran
| |
Collapse
|
19
|
Causal involvement of the left angular gyrus in higher functions as revealed by transcranial magnetic stimulation: a systematic review. Brain Struct Funct 2023; 228:169-196. [PMID: 36260126 DOI: 10.1007/s00429-022-02576-w] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2022] [Accepted: 09/21/2022] [Indexed: 01/07/2023]
Abstract
Transcranial magnetic stimulation (TMS) is a non-invasive technique that can transiently interfere with local cortical functioning, thus enabling inferences of causal left AG involvement in higher functions from experimentation with healthy participants. Here, we examine 35 studies that measure behavioural outcomes soon after or during targeting TMS to the left AG, by design and as documented by individual magnetic resonance images, in healthy adult participants. The reviewed evidence suggests a specific causal involvement of the left AG in a wide range of tasks involving language, memory, number processing, visuospatial attention, body awareness and motor planning functions. These core findings are particularly valuable to inform theoretical models of the left AG role(s) in higher functions, due to the anatomical specificity afforded by the selected studies and the complementarity of TMS to different methods of investigation. In particular, the variety of the operations within and between functions in which the left AG appears to be causally involved poses a formidable challenge to any attempts to identify a single computational process subserved by the left AG (as opposed to just outlining a broad type of functional contribution) that could apply across thematic areas. We conclude by highlighting directions for improvement in future experimentation with TMS, in order to strengthen the available evidence, while taking into account the anatomical heterogeneity of this brain region.
Collapse
|
20
|
The role of the angular gyrus in semantic cognition: a synthesis of five functional neuroimaging studies. Brain Struct Funct 2023; 228:273-291. [PMID: 35476027 DOI: 10.1007/s00429-022-02493-y] [Citation(s) in RCA: 17] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2021] [Accepted: 04/04/2022] [Indexed: 01/07/2023]
Abstract
Semantic knowledge is central to human cognition. The angular gyrus (AG) is widely considered a key brain region for semantic cognition. However, the role of the AG in semantic processing is controversial. Key controversies concern response polarity (activation vs. deactivation) and its relation to task difficulty, lateralization (left vs. right AG), and functional-anatomical subdivision (PGa vs. PGp subregions). Here, we combined the fMRI data of five studies on semantic processing (n = 172) and analyzed the response profiles from the same anatomical regions-of-interest for left and right PGa and PGp. We found that the AG was consistently deactivated during non-semantic conditions, whereas response polarity during semantic conditions was inconsistent. However, the AG consistently showed relative response differences between semantic and non-semantic conditions, and between different semantic conditions. A combined analysis across all studies revealed that AG responses could be best explained by separable effects of task difficulty and semantic processing demand. Task difficulty effects were stronger in PGa than PGp, regardless of hemisphere. Semantic effects were stronger in left than right AG, regardless of subregion. These results suggest that the AG is engaged in both domain-general task-difficulty-related processes and domain-specific semantic processes. In semantic processing, we propose that left AG acts as a "multimodal convergence zone" that binds different semantic features associated with the same concept, enabling efficient access to task-relevant features.
Collapse
|
21
|
MacGregor LJ, Gilbert RA, Balewski Z, Mitchell DJ, Erzinçlioğlu SW, Rodd JM, Duncan J, Fedorenko E, Davis MH. Causal Contributions of the Domain-General (Multiple Demand) and the Language-Selective Brain Networks to Perceptual and Semantic Challenges in Speech Comprehension. NEUROBIOLOGY OF LANGUAGE (CAMBRIDGE, MASS.) 2022; 3:665-698. [PMID: 36742011 PMCID: PMC9893226 DOI: 10.1162/nol_a_00081] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Accepted: 09/07/2022] [Indexed: 06/18/2023]
Abstract
Listening to spoken language engages domain-general multiple demand (MD; frontoparietal) regions of the human brain, in addition to domain-selective (frontotemporal) language regions, particularly when comprehension is challenging. However, there is limited evidence that the MD network makes a functional contribution to core aspects of understanding language. In a behavioural study of volunteers (n = 19) with chronic brain lesions, but without aphasia, we assessed the causal role of these networks in perceiving, comprehending, and adapting to spoken sentences made more challenging by acoustic-degradation or lexico-semantic ambiguity. We measured perception of and adaptation to acoustically degraded (noise-vocoded) sentences with a word report task before and after training. Participants with greater damage to MD but not language regions required more vocoder channels to achieve 50% word report, indicating impaired perception. Perception improved following training, reflecting adaptation to acoustic degradation, but adaptation was unrelated to lesion location or extent. Comprehension of spoken sentences with semantically ambiguous words was measured with a sentence coherence judgement task. Accuracy was high and unaffected by lesion location or extent. Adaptation to semantic ambiguity was measured in a subsequent word association task, which showed that availability of lower-frequency meanings of ambiguous words increased following their comprehension (word-meaning priming). Word-meaning priming was reduced for participants with greater damage to language but not MD regions. Language and MD networks make dissociable contributions to challenging speech comprehension: Using recent experience to update word meaning preferences depends on language-selective regions, whereas the domain-general MD network plays a causal role in reporting words from degraded speech.
Collapse
Affiliation(s)
- Lucy J. MacGregor
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Rebecca A. Gilbert
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Zuzanna Balewski
- Helen Wills Neuroscience Institute, University of California, Berkeley, Berkeley, CA
| | - Daniel J. Mitchell
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | | | - Jennifer M. Rodd
- Psychology and Language Sciences, University College London, London, UK
| | - John Duncan
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| | - Evelina Fedorenko
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA
- McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA
- Program in Speech and Hearing Bioscience and Technology, Harvard University, Cambridge, MA
| | - Matthew H. Davis
- MRC Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, UK
| |
Collapse
|
22
|
Scott SK, Jasmin K. Rostro-caudal networks for sound processing in the primate brain. Front Neurosci 2022; 16:1076374. [PMID: 36590301 PMCID: PMC9797816 DOI: 10.3389/fnins.2022.1076374] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Accepted: 11/28/2022] [Indexed: 12/23/2022] Open
Abstract
Sound is processed in primate brains along anatomically and functionally distinct streams: this pattern can be seen in both human and non-human primates. We have previously proposed a general auditory processing framework in which these different perceptual profiles are associated with different computational characteristics. In this paper we consider how recent work supports our framework.
Collapse
Affiliation(s)
- Sophie K. Scott
- Institute of Cognitive Neuroscience, University College London, London, United Kingdom,*Correspondence: Sophie K. Scott,
| | - Kyle Jasmin
- Department of Psychology, Royal Holloway, University of London, Egham, United Kingdom,Kyle Jasmin,
| |
Collapse
|
23
|
Wang S, Planton S, Chanoine V, Sein J, Anton JL, Nazarian B, Dubarry AS, Pallier C, Pattamadilok C. Graph theoretical analysis reveals the functional role of the left ventral occipito-temporal cortex in speech processing. Sci Rep 2022; 12:20028. [PMID: 36414688 PMCID: PMC9681757 DOI: 10.1038/s41598-022-24056-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2022] [Accepted: 11/09/2022] [Indexed: 11/23/2022] Open
Abstract
The left ventral occipito-temporal cortex (left-vOT) plays a key role in reading. Interestingly, the area also responds to speech input, suggesting that it may have other functions beyond written word recognition. Here, we adopt graph theoretical analysis to investigate the left-vOT's functional role in the whole-brain network while participants process spoken sentences in different contexts. Overall, different connectivity measures indicate that the left-vOT acts as an interface enabling the communication between distributed brain regions and sub-networks. During simple speech perception, the left-vOT is systematically part of the visual network and contributes to the communication between neighboring areas, remote areas, and sub-networks, by acting as a local bridge, a global bridge, and a connector, respectively. However, when speech comprehension is explicitly required, the specific functional role of the area and the sub-network to which the left-vOT belongs change and vary with the quality of speech signal and task difficulty. These connectivity patterns provide insightful information on the contribution of the left-vOT in various contexts of language processing beyond its role in reading. They advance our general understanding of the neural mechanisms underlying the flexibility of the language network that adjusts itself according to the processing context.
Collapse
Affiliation(s)
- Shuai Wang
- grid.462776.60000 0001 2206 2382Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France ,grid.5399.60000 0001 2176 4817Aix Marseille Univ, Institute of Language, Communication and the Brain, Aix-en-Provence, France
| | - Samuel Planton
- grid.462776.60000 0001 2206 2382Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France ,grid.7429.80000000121866389Cognitive Neuroimaging Unit, INSERM, CEA, CNRS, Université Paris-Saclay, NeuroSpin Center, Gif/Yvette, France
| | - Valérie Chanoine
- grid.462776.60000 0001 2206 2382Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France ,grid.5399.60000 0001 2176 4817Aix Marseille Univ, Institute of Language, Communication and the Brain, Aix-en-Provence, France
| | - Julien Sein
- grid.462486.a0000 0004 4650 2882Aix Marseille Univ, CNRS, Centre IRM-INT@CERIMED, Institut de Neurosciences de la Timone, UMR 7289 Marseille, France
| | - Jean-Luc Anton
- grid.462486.a0000 0004 4650 2882Aix Marseille Univ, CNRS, Centre IRM-INT@CERIMED, Institut de Neurosciences de la Timone, UMR 7289 Marseille, France
| | - Bruno Nazarian
- grid.462486.a0000 0004 4650 2882Aix Marseille Univ, CNRS, Centre IRM-INT@CERIMED, Institut de Neurosciences de la Timone, UMR 7289 Marseille, France
| | - Anne-Sophie Dubarry
- grid.462776.60000 0001 2206 2382Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France ,grid.4444.00000 0001 2112 9282 Aix Marseille Univ, CNRS, LNC, Marseille, France
| | - Christophe Pallier
- grid.7429.80000000121866389Cognitive Neuroimaging Unit, INSERM, CEA, CNRS, Université Paris-Saclay, NeuroSpin Center, Gif/Yvette, France
| | - Chotiga Pattamadilok
- grid.462776.60000 0001 2206 2382Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France
| |
Collapse
|
24
|
Qiao Y, Zhu M, Sun W, Sun Y, Guo H, Shang Y. Intrinsic brain activity reorganization contributes to long-term compensation of higher-order hearing abilities in single-sided deafness. Front Neurosci 2022; 16:935834. [PMID: 36090279 PMCID: PMC9453152 DOI: 10.3389/fnins.2022.935834] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Accepted: 08/01/2022] [Indexed: 11/24/2022] Open
Abstract
Single-sided deafness (SSD) is an extreme case of partial hearing deprivation and results in a significant decline in higher-order hearing abilities, including sound localization and speech-in-noise recognition. Clinical studies have reported that patients with SSD recover from these higher-order hearing abilities to some extent over time. Neuroimaging studies have observed extensive brain functional plasticity in patients with SSD. However, studies investigating the role of plasticity in functional compensation, particularly those investigating the relationship between intrinsic brain activity alterations and higher-order hearing abilities, are still limited. In this study, we used resting-state functional MRI to investigate intrinsic brain activity, measured by the amplitude of low-frequency fluctuation (ALFF), in 19 patients with left SSD, 17 patients with right SSD, and 21 normal hearing controls (NHs). All patients with SSD had durations of deafness longer than 2 years. Decreased ALFF values in the bilateral precuneus (PCUN), lingual gyrus, and left middle frontal gyrus were observed in patients with SSD compared with the values of NHs. Longer durations of deafness were correlated with better hearing abilities, as well as higher ALFF values in the left inferior parietal lobule, the angular gyrus, the middle occipital gyrus, the bilateral PCUN, and the posterior cingulate gyrus. Moreover, we observed a generally consistent trend of correlation between ALFF values and higher-order hearing abilities in specific brain areas in patients with SSD. That is, better abilities were correlated with lower ALFF values in the frontal regions and higher ALFF values in the PCUN and surrounding parietal-occipital areas. Furthermore, mediation analysis revealed that the ALFF values in the PCUN were a significant mediator of the relationship between the duration of deafness and higher-order hearing abilities. Our study reveals significant plasticity of intrinsic brain activity in patients with SSD and suggests that reorganization of intrinsic brain activity may be one of the compensatory mechanisms that facilitate improvement in higher-order hearing abilities in these patients over time.
Collapse
Affiliation(s)
- Yufei Qiao
- Department of Otorhinolaryngology, Peking Union Medical College Hospital, Beijing, China
| | - Min Zhu
- Department of Otorhinolaryngology, Peking Union Medical College Hospital, Beijing, China
| | - Wen Sun
- Department of Otorhinolaryngology, Peking Union Medical College Hospital, Beijing, China
| | - Yang Sun
- School of Educational Science, Shenyang Normal University, Shengyang, China
| | - Hua Guo
- Department of Biomedical Engineering, Center for Biomedical Imaging Research, School of Medicine, Tsinghua University, Beijing, China
| | - Yingying Shang
- Department of Otorhinolaryngology, Peking Union Medical College Hospital, Beijing, China
- *Correspondence: Yingying Shang
| |
Collapse
|
25
|
Zhou X, Burg E, Kan A, Litovsky RY. Investigating effortful speech perception using fNIRS and pupillometry measures. CURRENT RESEARCH IN NEUROBIOLOGY 2022; 3:100052. [PMID: 36518346 PMCID: PMC9743070 DOI: 10.1016/j.crneur.2022.100052] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2021] [Revised: 05/12/2022] [Accepted: 08/12/2022] [Indexed: 10/15/2022] Open
Abstract
The current study examined the neural mechanisms for mental effort and its correlation to speech perception using functional near-infrared spectroscopy (fNIRS) in listeners with normal hearing (NH). Data were collected while participants listened and responded to unprocessed and degraded sentences, where words were presented in grammatically correct or shuffled order. Effortful listening and task difficulty due to stimulus manipulations was confirmed using a subjective questionnaire and a well-established objective measure of mental effort - pupillometry. fNIRS measures focused on cortical responses in two a priori regions of interest, the left auditory cortex (AC) and lateral frontal cortex (LFC), which are closely related to auditory speech perception and listening effort, respectively. We examined the relations between the two objective measures and behavioral measures of speech perception (task performance) and task difficulty. Results demonstrated that changes in pupil dilation were positively correlated with the self-reported task difficulty levels and negatively correlated with the task performance scores. A significant and negative correlation between the two behavioral measures was also found. That is, as perceived task demands increased and task performance scores decreased, pupils dilated more. fNIRS measures (cerebral oxygenation) in the left AC and LFC were both negatively correlated with the self-reported task difficulty levels and positively correlated with task performance scores. These results suggest that pupillometry measures can indicate task demands and listening effort; whereas, fNIRS measures using a similar paradigm seem to reflect speech processing, but not effort.
Collapse
Affiliation(s)
- Xin Zhou
- Waisman Center, University of Wisconsin Madison, WI, USA
| | - Emily Burg
- Waisman Center, University of Wisconsin Madison, WI, USA
- Department of Communication Science and Disorders, University of Wisconsin Madison, WI, USA
| | - Alan Kan
- School of Engineering, Macquarie University, Sydney, NSW, Australia
| | - Ruth Y Litovsky
- Waisman Center, University of Wisconsin Madison, WI, USA
- Department of Communication Science and Disorders, University of Wisconsin Madison, WI, USA
| |
Collapse
|
26
|
Age-related differences in the neural network interactions underlying the predictability gain. Cortex 2022; 154:269-286. [DOI: 10.1016/j.cortex.2022.05.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2021] [Revised: 03/30/2022] [Accepted: 05/03/2022] [Indexed: 11/20/2022]
|
27
|
Hauswald A, Keitel A, Chen Y, Rösch S, Weisz N. Degradation levels of continuous speech affect neural speech tracking and alpha power differently. Eur J Neurosci 2022; 55:3288-3302. [PMID: 32687616 PMCID: PMC9540197 DOI: 10.1111/ejn.14912] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2020] [Revised: 07/12/2020] [Accepted: 07/13/2020] [Indexed: 11/26/2022]
Abstract
Making sense of a poor auditory signal can pose a challenge. Previous attempts to quantify speech intelligibility in neural terms have usually focused on one of two measures, namely low-frequency speech-brain synchronization or alpha power modulations. However, reports have been mixed concerning the modulation of these measures, an issue aggravated by the fact that they have normally been studied separately. We present two MEG studies analyzing both measures. In study 1, participants listened to unimodal auditory speech with three different levels of degradation (original, 7-channel and 3-channel vocoding). Intelligibility declined with declining clarity, but speech was still intelligible to some extent even for the lowest clarity level (3-channel vocoding). Low-frequency (1-7 Hz) speech tracking suggested a U-shaped relationship with strongest effects for the medium-degraded speech (7-channel) in bilateral auditory and left frontal regions. To follow up on this finding, we implemented three additional vocoding levels (5-channel, 2-channel and 1-channel) in a second MEG study. Using this wider range of degradation, the speech-brain synchronization showed a similar pattern as in study 1, but further showed that when speech becomes unintelligible, synchronization declines again. The relationship differed for alpha power, which continued to decrease across vocoding levels reaching a floor effect for 5-channel vocoding. Predicting subjective intelligibility based on models either combining both measures or each measure alone showed superiority of the combined model. Our findings underline that speech tracking and alpha power are modified differently by the degree of degradation of continuous speech but together contribute to the subjective speech understanding.
Collapse
Affiliation(s)
- Anne Hauswald
- Center of Cognitive NeuroscienceUniversity of SalzburgSalzburgAustria
- Department of PsychologyUniversity of SalzburgSalzburgAustria
| | - Anne Keitel
- Psychology, School of Social SciencesUniversity of DundeeDundeeUK
- Centre for Cognitive NeuroimagingUniversity of GlasgowGlasgowUK
| | - Ya‐Ping Chen
- Center of Cognitive NeuroscienceUniversity of SalzburgSalzburgAustria
- Department of PsychologyUniversity of SalzburgSalzburgAustria
| | - Sebastian Rösch
- Department of OtorhinolaryngologyParacelsus Medical UniversitySalzburgAustria
| | - Nathan Weisz
- Center of Cognitive NeuroscienceUniversity of SalzburgSalzburgAustria
- Department of PsychologyUniversity of SalzburgSalzburgAustria
| |
Collapse
|
28
|
Irsik VC, Johnsrude IS, Herrmann B. Age-related deficits in dip-listening evident for isolated sentences but not for spoken stories. Sci Rep 2022; 12:5898. [PMID: 35393472 PMCID: PMC8991280 DOI: 10.1038/s41598-022-09805-6] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2021] [Accepted: 03/23/2022] [Indexed: 12/03/2022] Open
Abstract
Fluctuating background sounds facilitate speech intelligibility by providing speech ‘glimpses’ (masking release). Older adults benefit less from glimpses, but masking release is typically investigated using isolated sentences. Recent work indicates that using engaging, continuous speech materials (e.g., spoken stories) may qualitatively alter speech-in-noise listening. Moreover, neural sensitivity to different amplitude envelope profiles (ramped, damped) changes with age, but whether this affects speech listening is unknown. In three online experiments, we investigate how masking release in younger and older adults differs for masked sentences and stories, and how speech intelligibility varies with masker amplitude profile. Intelligibility was generally greater for damped than ramped maskers. Masking release was reduced in older relative to younger adults for disconnected sentences, and stories with a randomized sentence order. Critically, when listening to stories with an engaging and coherent narrative, older adults demonstrated equal or greater masking release compared to younger adults. Older adults thus appear to benefit from ‘glimpses’ as much as, or more than, younger adults when the speech they are listening to follows a coherent topical thread. Our results highlight the importance of cognitive and motivational factors for speech understanding, and suggest that previous work may have underestimated speech-listening abilities in older adults.
Collapse
Affiliation(s)
- Vanessa C Irsik
- Department of Psychology & The Brain and Mind Institute, The University of Western Ontario, London, ON, N6A 3K7, Canada.
| | - Ingrid S Johnsrude
- Department of Psychology & The Brain and Mind Institute, The University of Western Ontario, London, ON, N6A 3K7, Canada.,School of Communication and Speech Disorders, The University of Western Ontario, London, ON, N6A 5B7, Canada
| | - Björn Herrmann
- Department of Psychology & The Brain and Mind Institute, The University of Western Ontario, London, ON, N6A 3K7, Canada.,Rotman Research Institute, Baycrest, Toronto, ON, M6A 2E1, Canada.,Department of Psychology, University of Toronto, Toronto, ON, M5S 1A1, Canada
| |
Collapse
|
29
|
Bsharat-Maalouf D, Karawani H. Bilinguals' speech perception in noise: Perceptual and neural associations. PLoS One 2022; 17:e0264282. [PMID: 35196339 PMCID: PMC8865662 DOI: 10.1371/journal.pone.0264282] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2021] [Accepted: 02/07/2022] [Indexed: 01/26/2023] Open
Abstract
The current study characterized subcortical speech sound processing among monolinguals and bilinguals in quiet and challenging listening conditions and examined the relation between subcortical neural processing and perceptual performance. A total of 59 normal-hearing adults, ages 19–35 years, participated in the study: 29 native Hebrew-speaking monolinguals and 30 Arabic-Hebrew-speaking bilinguals. Auditory brainstem responses to speech sounds were collected in a quiet condition and with background noise. The perception of words and sentences in quiet and background noise conditions was also examined to assess perceptual performance and to evaluate the perceptual-physiological relationship. Perceptual performance was tested among bilinguals in both languages (first language (L1-Arabic) and second language (L2-Hebrew)). The outcomes were similar between monolingual and bilingual groups in quiet. Noise, as expected, resulted in deterioration in perceptual and neural responses, which was reflected in lower accuracy in perceptual tasks compared to quiet, and in more prolonged latencies and diminished neural responses. However, a mixed picture was observed among bilinguals in perceptual and physiological outcomes in noise. In the perceptual measures, bilinguals were significantly less accurate than their monolingual counterparts. However, in neural responses, bilinguals demonstrated earlier peak latencies compared to monolinguals. Our results also showed that perceptual performance in noise was related to subcortical resilience to the disruption caused by background noise. Specifically, in noise, increased brainstem resistance (i.e., fewer changes in the fundamental frequency (F0) representations or fewer shifts in the neural timing) was related to better speech perception among bilinguals. Better perception in L1 in noise was correlated with fewer changes in F0 representations, and more accurate perception in L2 was related to minor shifts in auditory neural timing. This study delves into the importance of using neural brainstem responses to speech sounds to differentiate individuals with different language histories and to explain inter-subject variability in bilinguals’ perceptual abilities in daily life situations.
Collapse
Affiliation(s)
- Dana Bsharat-Maalouf
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
| | - Hanin Karawani
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
- * E-mail:
| |
Collapse
|
30
|
Irsik VC, Johnsrude IS, Herrmann B. Neural Activity during Story Listening Is Synchronized across Individuals Despite Acoustic Masking. J Cogn Neurosci 2022; 34:933-950. [PMID: 35258555 DOI: 10.1162/jocn_a_01842] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Older people with hearing problems often experience difficulties understanding speech in the presence of background sound. As a result, they may disengage in social situations, which has been associated with negative psychosocial health outcomes. Measuring listening (dis)engagement during challenging listening situations has received little attention thus far. We recruit young, normal-hearing human adults (both sexes) and investigate how speech intelligibility and engagement during naturalistic story listening is affected by the level of acoustic masking (12-talker babble) at different signal-to-noise ratios (SNRs). In Experiment 1, we observed that word-report scores were above 80% for all but the lowest SNR (-3 dB SNR) we tested, at which performance dropped to 54%. In Experiment 2, we calculated intersubject correlation (ISC) using EEG data to identify dynamic spatial patterns of shared neural activity evoked by the stories. ISC has been used as a neural measure of participants' engagement with naturalistic materials. Our results show that ISC was stable across all but the lowest SNRs, despite reduced speech intelligibility. Comparing ISC and intelligibility demonstrated that word-report performance declined more strongly with decreasing SNR compared to ISC. Our measure of neural engagement suggests that individuals remain engaged in story listening despite missing words because of background noise. Our work provides a potentially fruitful approach to investigate listener engagement with naturalistic, spoken stories that may be used to investigate (dis)engagement in older adults with hearing impairment.
Collapse
Affiliation(s)
| | | | - Björn Herrmann
- The University of Western Ontario.,Rotman Research Institute, Toronto, ON, Canada.,University of Toronto
| |
Collapse
|
31
|
Knipper M, Singer W, Schwabe K, Hagberg GE, Li Hegner Y, Rüttiger L, Braun C, Land R. Disturbed Balance of Inhibitory Signaling Links Hearing Loss and Cognition. Front Neural Circuits 2022; 15:785603. [PMID: 35069123 PMCID: PMC8770933 DOI: 10.3389/fncir.2021.785603] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Accepted: 12/08/2021] [Indexed: 12/19/2022] Open
Abstract
Neuronal hyperexcitability in the central auditory pathway linked to reduced inhibitory activity is associated with numerous forms of hearing loss, including noise damage, age-dependent hearing loss, and deafness, as well as tinnitus or auditory processing deficits in autism spectrum disorder (ASD). In most cases, the reduced central inhibitory activity and the accompanying hyperexcitability are interpreted as an active compensatory response to the absence of synaptic activity, linked to increased central neural gain control (increased output activity relative to reduced input). We here suggest that hyperexcitability also could be related to an immaturity or impairment of tonic inhibitory strength that typically develops in an activity-dependent process in the ascending auditory pathway with auditory experience. In these cases, high-SR auditory nerve fibers, which are critical for the shortest latencies and lowest sound thresholds, may have either not matured (possibly in congenital deafness or autism) or are dysfunctional (possibly after sudden, stressful auditory trauma or age-dependent hearing loss linked with cognitive decline). Fast auditory processing deficits can occur despite maintained basal hearing. In that case, tonic inhibitory strength is reduced in ascending auditory nuclei, and fast inhibitory parvalbumin positive interneuron (PV-IN) dendrites are diminished in auditory and frontal brain regions. This leads to deficits in central neural gain control linked to hippocampal LTP/LTD deficiencies, cognitive deficits, and unbalanced extra-hypothalamic stress control. Under these conditions, a diminished inhibitory strength may weaken local neuronal coupling to homeostatic vascular responses required for the metabolic support of auditory adjustment processes. We emphasize the need to distinguish these two states of excitatory/inhibitory imbalance in hearing disorders: (i) Under conditions of preserved fast auditory processing and sustained tonic inhibitory strength, an excitatory/inhibitory imbalance following auditory deprivation can maintain precise hearing through a memory linked, transient disinhibition that leads to enhanced spiking fidelity (central neural gain⇑) (ii) Under conditions of critically diminished fast auditory processing and reduced tonic inhibitory strength, hyperexcitability can be part of an increased synchronization over a broader frequency range, linked to reduced spiking reliability (central neural gain⇓). This latter stage mutually reinforces diminished metabolic support for auditory adjustment processes, increasing the risks for canonical dementia syndromes.
Collapse
Affiliation(s)
- Marlies Knipper
- Department of Otolaryngology, Head and Neck Surgery, Tübingen Hearing Research Center (THRC), Molecular Physiology of Hearing, University of Tübingen, Tübingen, Germany
- *Correspondence: Marlies Knipper,
| | - Wibke Singer
- Department of Otolaryngology, Head and Neck Surgery, Tübingen Hearing Research Center (THRC), Molecular Physiology of Hearing, University of Tübingen, Tübingen, Germany
| | - Kerstin Schwabe
- Experimental Neurosurgery, Department of Neurosurgery, Hannover Medical School, Hanover, Germany
| | - Gisela E. Hagberg
- Department of Biomedical Magnetic Resonance, University Hospital Tübingen (UKT), Tübingen, Germany
- High-Field Magnetic Resonance, Max Planck Institute for Biological Cybernetics, Tübingen, Germany
| | - Yiwen Li Hegner
- MEG Center, University of Tübingen, Tübingen, Germany
- Center of Neurology, Hertie-Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
| | - Lukas Rüttiger
- Department of Otolaryngology, Head and Neck Surgery, Tübingen Hearing Research Center (THRC), Molecular Physiology of Hearing, University of Tübingen, Tübingen, Germany
| | - Christoph Braun
- MEG Center, University of Tübingen, Tübingen, Germany
- Center of Neurology, Hertie-Institute for Clinical Brain Research, University of Tübingen, Tübingen, Germany
| | - Rüdiger Land
- Department of Experimental Otology, Institute for Audioneurotechnology, Hannover Medical School, Hanover, Germany
| |
Collapse
|
32
|
Al-Zubaidi A, Bräuer S, Holdgraf CR, Schepers IM, Rieger JW. OUP accepted manuscript. Cereb Cortex Commun 2022; 3:tgac007. [PMID: 35281216 PMCID: PMC8914075 DOI: 10.1093/texcom/tgac007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2020] [Revised: 01/26/2022] [Accepted: 01/29/2022] [Indexed: 11/14/2022] Open
Affiliation(s)
- Arkan Al-Zubaidi
- Applied Neurocognitive Psychology Lab and Cluster of Excellence Hearing4all, Oldenburg University, Oldenburg, Germany
- Research Center Neurosensory Science, Oldenburg University, 26129 Oldenburg, Germany
| | - Susann Bräuer
- Applied Neurocognitive Psychology Lab and Cluster of Excellence Hearing4all, Oldenburg University, Oldenburg, Germany
| | - Chris R Holdgraf
- Department of Statistics, UC Berkeley, Berkeley, CA 94720, USA
- International Interactive Computing Collaboration
| | - Inga M Schepers
- Applied Neurocognitive Psychology Lab and Cluster of Excellence Hearing4all, Oldenburg University, Oldenburg, Germany
| | - Jochem W Rieger
- Corresponding author: Department of Psychology, Faculty VI, Oldenburg University, 26129 Oldenburg, Germany.
| |
Collapse
|
33
|
Agmon G, Yahav PHS, Ben-Shachar M, Golumbic EZ. Attention to Speech: Mapping Distributed and Selective Attention Systems. Cereb Cortex 2021; 32:3763-3776. [PMID: 34875678 DOI: 10.1093/cercor/bhab446] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2021] [Revised: 11/02/2021] [Accepted: 11/03/2021] [Indexed: 11/14/2022] Open
Abstract
When faced with situations where many people talk at once, individuals can employ different listening strategies to deal with the cacophony of speech sounds and to achieve different goals. In this fMRI study, we investigated how the pattern of neural activity is affected by the type of attention applied to speech in a simulated "cocktail party." Specifically, we compared brain activation patterns when listeners "attended selectively" to only one speaker and ignored all others, versus when they "distributed their attention" and followed several concurrent speakers. Conjunction analysis revealed a highly overlapping network of regions activated for both types of attention, including auditory association cortex (bilateral STG/STS) and frontoparietal regions related to speech processing and attention (bilateral IFG/insula, right MFG, left IPS). Activity within nodes of this network, though, was modulated by the type of attention required as well as the number of competing speakers. Auditory and speech-processing regions exhibited higher activity during distributed attention, whereas frontoparietal regions were activated more strongly during selective attention. These results suggest a common "attention to speech" network, which provides the computational infrastructure to deal effectively with multi-speaker input, but with sufficient flexibility to implement different prioritization strategies and to adapt to different listener goals.
Collapse
Affiliation(s)
- Galit Agmon
- Gonda Multidisciplinary Brain Research Center, Bar-Ilan University, Ramat-Gan 5290002, Israel
| | - Paz Har-Shai Yahav
- Gonda Multidisciplinary Brain Research Center, Bar-Ilan University, Ramat-Gan 5290002, Israel
| | - Michal Ben-Shachar
- Gonda Multidisciplinary Brain Research Center, Bar-Ilan University, Ramat-Gan 5290002, Israel.,Department of English Literature and Linguistics, Bar-Ilan University, Ramat-Gan 5290002, Israel
| | - Elana Zion Golumbic
- Gonda Multidisciplinary Brain Research Center, Bar-Ilan University, Ramat-Gan 5290002, Israel
| |
Collapse
|
34
|
Qi T, Schaadt G, Friederici AD. Associated functional network development and language abilities in children. Neuroimage 2021; 242:118452. [PMID: 34358655 PMCID: PMC8463838 DOI: 10.1016/j.neuroimage.2021.118452] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2021] [Revised: 07/14/2021] [Accepted: 08/03/2021] [Indexed: 11/26/2022] Open
Abstract
During childhood, the brain is gradually converging to the efficient functional architecture observed in adults. How the brain's functional architecture evolves with age, particularly in young children, is however, not well understood. We examined the functional connectivity of the core language regions, in association with cortical growth and language abilities, in 175 young children in the age range of 4 to 9 years. We analyzed the brain's developmental changes using resting-state functional and T1-weighted structural magnetic resonance imaging data. The results showed increased functional connectivity strength with age between the pars triangularis of the left inferior frontal gyrus and left temporoparietal regions (cohen's d = 0.54, CI: 0.24 - 0.84), associated with children's language abilities. Stronger functional connectivity between bilateral prefrontal and temporoparietal regions was associated with better language abilities regardless of age. In addition, the stronger functional connectivity between the left inferior frontal and temporoparietal regions was associated with larger surface area and thinner cortical thickness in these regions, which in turn was associated with superior language abilities. Thus, using functional and structural brain indices, coupled with behavioral measures, we elucidate the association of functional language network development, language ability, and cortical growth, thereby adding to our understanding of the neural basis of language acquisition in young children.
Collapse
Affiliation(s)
- Ting Qi
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.
| | - Gesa Schaadt
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany; Department of Education and Psychology, Free University of Berlin, Berlin, Germany
| | - Angela D Friederici
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| |
Collapse
|
35
|
Defenderfer J, Forbes S, Wijeakumar S, Hedrick M, Plyler P, Buss AT. Frontotemporal activation differs between perception of simulated cochlear implant speech and speech in background noise: An image-based fNIRS study. Neuroimage 2021; 240:118385. [PMID: 34256138 PMCID: PMC8503862 DOI: 10.1016/j.neuroimage.2021.118385] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2021] [Revised: 06/10/2021] [Accepted: 07/09/2021] [Indexed: 10/27/2022] Open
Abstract
In this study we used functional near-infrared spectroscopy (fNIRS) to investigate neural responses in normal-hearing adults as a function of speech recognition accuracy, intelligibility of the speech stimulus, and the manner in which speech is distorted. Participants listened to sentences and reported aloud what they heard. Speech quality was distorted artificially by vocoding (simulated cochlear implant speech) or naturally by adding background noise. Each type of distortion included high and low-intelligibility conditions. Sentences in quiet were used as baseline comparison. fNIRS data were analyzed using a newly developed image reconstruction approach. First, elevated cortical responses in the middle temporal gyrus (MTG) and middle frontal gyrus (MFG) were associated with speech recognition during the low-intelligibility conditions. Second, activation in the MTG was associated with recognition of vocoded speech with low intelligibility, whereas MFG activity was largely driven by recognition of speech in background noise, suggesting that the cortical response varies as a function of distortion type. Lastly, an accuracy effect in the MFG demonstrated significantly higher activation during correct perception relative to incorrect perception of speech. These results suggest that normal-hearing adults (i.e., untrained listeners of vocoded stimuli) do not exploit the same attentional mechanisms of the frontal cortex used to resolve naturally degraded speech and may instead rely on segmental and phonetic analyses in the temporal lobe to discriminate vocoded speech.
Collapse
Affiliation(s)
- Jessica Defenderfer
- Speech and Hearing Science, University of Tennessee Health Science Center, Knoxville, TN, United States.
| | - Samuel Forbes
- Psychology, University of East Anglia, Norwich, England.
| | | | - Mark Hedrick
- Speech and Hearing Science, University of Tennessee Health Science Center, Knoxville, TN, United States.
| | - Patrick Plyler
- Speech and Hearing Science, University of Tennessee Health Science Center, Knoxville, TN, United States.
| | - Aaron T Buss
- Psychology, University of Tennessee, Knoxville, TN, United States.
| |
Collapse
|
36
|
Reduced Semantic Context and Signal-to-Noise Ratio Increase Listening Effort As Measured Using Functional Near-Infrared Spectroscopy. Ear Hear 2021; 43:836-848. [PMID: 34623112 DOI: 10.1097/aud.0000000000001137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
OBJECTIVES Understanding speech-in-noise can be highly effortful. Decreasing the signal-to-noise ratio (SNR) of speech increases listening effort, but it is relatively unclear if decreasing the level of semantic context does as well. The current study used functional near-infrared spectroscopy to evaluate two primary hypotheses: (1) listening effort (operationalized as oxygenation of the left lateral PFC) increases as the SNR decreases and (2) listening effort increases as context decreases. DESIGN Twenty-eight younger adults with normal hearing completed the Revised Speech Perception in Noise Test, in which they listened to sentences and reported the final word. These sentences either had an easy SNR (+4 dB) or a hard SNR (-2 dB), and were either low in semantic context (e.g., "Tom could have thought about the sport") or high in context (e.g., "She had to vacuum the rug"). PFC oxygenation was measured throughout using functional near-infrared spectroscopy. RESULTS Accuracy on the Revised Speech Perception in Noise Test was worse when the SNR was hard than when it was easy, and worse for sentences low in semantic context than high in context. Similarly, oxygenation across the entire PFC (including the left lateral PFC) was greater when the SNR was hard, and left lateral PFC oxygenation was greater when context was low. CONCLUSIONS These results suggest that activation of the left lateral PFC (interpreted here as reflecting listening effort) increases to compensate for acoustic and linguistic challenges. This may reflect the increased engagement of domain-general and domain-specific processes subserved by the dorsolateral prefrontal cortex (e.g., cognitive control) and inferior frontal gyrus (e.g., predicting the sensory consequences of articulatory gestures), respectively.
Collapse
|
37
|
Bhandari P, Demberg V, Kray J. Semantic Predictability Facilitates Comprehension of Degraded Speech in a Graded Manner. Front Psychol 2021; 12:714485. [PMID: 34566795 PMCID: PMC8459870 DOI: 10.3389/fpsyg.2021.714485] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2021] [Accepted: 08/06/2021] [Indexed: 01/02/2023] Open
Abstract
Previous studies have shown that at moderate levels of spectral degradation, semantic predictability facilitates language comprehension. It is argued that when speech is degraded, listeners have narrowed expectations about the sentence endings; i.e., semantic prediction may be limited to only most highly predictable sentence completions. The main objectives of this study were to (i) examine whether listeners form narrowed expectations or whether they form predictions across a wide range of probable sentence endings, (ii) assess whether the facilitatory effect of semantic predictability is modulated by perceptual adaptation to degraded speech, and (iii) use and establish a sensitive metric for the measurement of language comprehension. For this, we created 360 German Subject-Verb-Object sentences that varied in semantic predictability of a sentence-final target word in a graded manner (high, medium, and low) and levels of spectral degradation (1, 4, 6, and 8 channels noise-vocoding). These sentences were presented auditorily to two groups: One group (n =48) performed a listening task in an unpredictable channel context in which the degraded speech levels were randomized, while the other group (n =50) performed the task in a predictable channel context in which the degraded speech levels were blocked. The results showed that at 4 channels noise-vocoding, response accuracy was higher in high-predictability sentences than in the medium-predictability sentences, which in turn was higher than in the low-predictability sentences. This suggests that, in contrast to the narrowed expectations view, comprehension of moderately degraded speech, ranging from low- to high- including medium-predictability sentences, is facilitated in a graded manner; listeners probabilistically preactivate upcoming words from a wide range of semantic space, not limiting only to highly probable sentence endings. Additionally, in both channel contexts, we did not observe learning effects; i.e., response accuracy did not increase over the course of experiment, and response accuracy was higher in the predictable than in the unpredictable channel context. We speculate from these observations that when there is no trial-by-trial variation of the levels of speech degradation, listeners adapt to speech quality at a long timescale; however, when there is a trial-by-trial variation of the high-level semantic feature (e.g., sentence predictability), listeners do not adapt to low-level perceptual property (e.g., speech quality) at a short timescale.
Collapse
Affiliation(s)
- Pratik Bhandari
- Department of Psychology, Saarland University, Saarbrücken, Germany
- Department of Language Science and Technology, Saarland University, Saarbrücken, Germany
| | - Vera Demberg
- Department of Language Science and Technology, Saarland University, Saarbrücken, Germany
- Department of Computer Science, Saarland University, Saarbrücken, Germany
| | - Jutta Kray
- Department of Psychology, Saarland University, Saarbrücken, Germany
| |
Collapse
|
38
|
White BE, Langdon C. The cortical organization of listening effort: New insight from functional near-infrared spectroscopy. Neuroimage 2021; 240:118324. [PMID: 34217787 DOI: 10.1016/j.neuroimage.2021.118324] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2021] [Revised: 06/17/2021] [Accepted: 06/28/2021] [Indexed: 10/21/2022] Open
Abstract
Everyday challenges impact our ability to hear and comprehend spoken language with ease, such as accented speech (source factors), spectral degradation (transmission factors), complex or unfamiliar language use (message factors), and predictability (context factors). Auditory degradation and linguistic complexity in the brain and behavior have been well investigated, and several computational models have emerged. The work here provides a novel test of the hypotheses that listening effort is partially reliant on higher cognitive auditory attention and working memory mechanisms in the frontal lobe, and partially reliant on hierarchical linguistic computation in the brain's left hemisphere. We specifically hypothesize that these models are robust and can be applied in ecologically relevant and coarse-grain contexts that rigorously control for acoustic and linguistic listening challenges. Using functional near-infrared spectroscopy during an auditory plausibility judgment task, we show the hierarchical cortical organization for listening effort in the frontal and left temporal-parietal brain regions. In response to increasing levels of cognitive demand, we found (i) poorer comprehension, (ii) slower reaction times, (iii) increasing levels of perceived mental effort, (iv) increasing levels of brain activity in the prefrontal cortex, (v) hierarchical modulation of core language processing regions that reflect increasingly higher-order auditory-linguistic processing, and (vi) a correlation between participants' mental effort ratings and their performance on the task. Our results demonstrate that listening effort is partly reliant on higher cognitive auditory attention and working memory mechanisms in the frontal lobe and partly reliant on hierarchical linguistic computation in the brain's left hemisphere. Further, listening effort is driven by a voluntary, motivation-based attention system for which our results validate the use of a single-item post-task questionnaire for measuring perceived levels of mental effort and predicting listening performance. We anticipate our study to be a starting point for more sophisticated models of listening effort and even cognitive neuroplasticity in hearing aid and cochlear implant users.
Collapse
Affiliation(s)
- Bradley E White
- Brain and Language Center for Neuroimaging, Gallaudet University, Washington, DC, USA.
| | - Clifton Langdon
- Department of Psychological Sciences, University of Connecticut, Storrs, CT, USA
| |
Collapse
|
39
|
Mechtenberg H, Xie X, Myers EB. Sentence predictability modulates cortical response to phonetic ambiguity. BRAIN AND LANGUAGE 2021; 218:104959. [PMID: 33930722 PMCID: PMC8513138 DOI: 10.1016/j.bandl.2021.104959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/13/2020] [Revised: 03/02/2021] [Accepted: 04/09/2021] [Indexed: 06/12/2023]
Abstract
Phonetic categories have undefined edges, such that individual tokens that belong to different speech sound categories may occupy the same region in acoustic space. In continuous speech, there are multiple sources of top-down information (e.g., lexical, semantic) that help to resolve the identity of an ambiguous phoneme. Of interest is how these top-down constraints interact with ambiguity at the phonetic level. In the current fMRI study, participants passively listened to sentences that varied in semantic predictability and in the amount of naturally-occurring phonetic competition. The left middle frontal gyrus, angular gyrus, and anterior inferior frontal gyrus were sensitive to both semantic predictability and the degree of phonetic competition. Notably, greater phonetic competition within non-predictive contexts resulted in a negatively-graded neural response. We suggest that uncertainty at the phonetic-acoustic level interacts with uncertainty at the semantic level-perhaps due to a failure of the network to construct a coherent meaning.
Collapse
Affiliation(s)
- Hannah Mechtenberg
- Department of Speech, Language, and Hearing Sciences, University of Connecticut, Storrs, Mansfield, CT 06269, USA.
| | - Xin Xie
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, NY 14627, USA.
| | - Emily B Myers
- Department of Speech, Language, and Hearing Sciences, University of Connecticut, Storrs, Mansfield, CT 06269, USA; Department of Psychological Sciences, University of Connecticut, Storrs, Mansfield, CT 06269, USA.
| |
Collapse
|
40
|
Dressing A, Kaller CP, Martin M, Nitschke K, Kuemmerer D, Beume LA, Schmidt CSM, Musso M, Urbach H, Rijntjes M, Weiller C. Anatomical correlates of recovery in apraxia: A longitudinal lesion-mapping study in stroke patients. Cortex 2021; 142:104-121. [PMID: 34265734 DOI: 10.1016/j.cortex.2021.06.001] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Revised: 05/07/2021] [Accepted: 06/01/2021] [Indexed: 11/24/2022]
Abstract
OBJECTIVE This study investigates the clinical course of recovery of apraxia after left-hemisphere stroke and the underlying neuroanatomical correlates for persisting or recovering deficits in relation to the major processing streams in the network for motor cognition. METHODS 90 patients were examined during the acute (4.74 ± 2.73 days) and chronic (14.3 ± 15.39 months) stage after left-hemisphere stroke for deficits in meaningless imitation, as well as production and conceptual errors in tool use pantomime. Lesion correlates for persisting or recovering deficits were analyzed with an extension of the non-parametric Brunner-Munzel rank-order test for multi-factorial designs (two-way repeated-measures ANOVA) using acute images. RESULTS Meaningless imitation and tool use production deficits persisted into the chronic stage. Conceptual errors in tool use pantomime showed an almost complete recovery. Imitation errors persisted after occipitotemporal and superior temporal lesions in the dorso-dorsal stream. Chronic pantomime production errors were related to the supramarginal gyrus, the key structure of the ventro-dorsal stream. More anterior lesions in the ventro-dorsal stream (ventral premotor cortex) were additionally associated with poor recovery of production errors in pantomime. Conceptual errors in pantomime after temporal and supramarginal gyrus lesions persisted into the chronic stage. However, they resolved completely when related to angular gyrus or insular lesions. CONCLUSION The diverging courses of recovery in different apraxia tasks can be related to different mechanisms. Critical lesions to key structures of the network or entrance areas of the processing streams lead to persisting deficits in the corresponding tasks. Contrary, lesions located outside the core network but inducing a temporary network dysfunction allow good recovery e.g., of conceptual errors in pantomime. The identification of lesion correlates for different long-term recovery patterns in apraxia might also allow early clinical prediction of the course of recovery.
Collapse
Affiliation(s)
- Andrea Dressing
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; BrainLinks-BrainTools Cluster of Excellence, University of Freiburg, Freiburg, Germany.
| | - Christoph P Kaller
- Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; BrainLinks-BrainTools Cluster of Excellence, University of Freiburg, Freiburg, Germany; Dept. of Neuroradiology, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Markus Martin
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; BrainLinks-BrainTools Cluster of Excellence, University of Freiburg, Freiburg, Germany
| | - Kai Nitschke
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Dorothee Kuemmerer
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Lena-A Beume
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Charlotte S M Schmidt
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Mariacristina Musso
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; BrainLinks-BrainTools Cluster of Excellence, University of Freiburg, Freiburg, Germany
| | - Horst Urbach
- Dept. of Neuroradiology, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Michel Rijntjes
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Cornelius Weiller
- Department of Neurology and Clinical Neuroscience, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; Freiburg Brain Imaging Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany; BrainLinks-BrainTools Cluster of Excellence, University of Freiburg, Freiburg, Germany
| |
Collapse
|
41
|
Guediche S, de Bruin A, Caballero-Gaudes C, Baart M, Samuel AG. Second-language word recognition in noise: Interdependent neuromodulatory effects of semantic context and crosslinguistic interactions driven by word form similarity. Neuroimage 2021; 237:118168. [PMID: 34000398 DOI: 10.1016/j.neuroimage.2021.118168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Revised: 05/05/2021] [Accepted: 05/12/2021] [Indexed: 11/17/2022] Open
Abstract
Spoken language comprehension is a fundamental component of our cognitive skills. We are quite proficient at deciphering words from the auditory input despite the fact that the speech we hear is often masked by noise such as background babble originating from talkers other than the one we are attending to. To perceive spoken language as intended, we rely on prior linguistic knowledge and context. Prior knowledge includes all sounds and words that are familiar to a listener and depends on linguistic experience. For bilinguals, the phonetic and lexical repertoire encompasses two languages, and the degree of overlap between word forms across languages affects the degree to which they influence one another during auditory word recognition. To support spoken word recognition, listeners often rely on semantic information (i.e., the words we hear are usually related in a meaningful way). Although the number of multilinguals across the globe is increasing, little is known about how crosslinguistic effects (i.e., word overlap) interact with semantic context and affect the flexible neural systems that support accurate word recognition. The current multi-echo functional magnetic resonance imaging (fMRI) study addresses this question by examining how prime-target word pair semantic relationships interact with the target word's form similarity (cognate status) to the translation equivalent in the dominant language (L1) during accurate word recognition of a non-dominant (L2) language. We tested 26 early-proficient Spanish-Basque (L1-L2) bilinguals. When L2 targets matching L1 translation-equivalent phonological word forms were preceded by unrelated semantic contexts that drive lexical competition, a flexible language control (fronto-parietal-subcortical) network was upregulated, whereas when they were preceded by related semantic contexts that reduce lexical competition, it was downregulated. We conclude that an interplay between semantic and crosslinguistic effects regulates flexible control mechanisms of speech processing to facilitate L2 word recognition, in noise.
Collapse
Affiliation(s)
- Sara Guediche
- Basque Center on Cognition Brain, and Language, Donostia-San Sebastian 20009, Spain.
| | | | | | - Martijn Baart
- Basque Center on Cognition Brain, and Language, Donostia-San Sebastian 20009, Spain; Department of Cognitive Neuropsychology, Tilburg University, P.O. Box 90153, 5000 LE Tilburg, the Netherlands
| | - Arthur G Samuel
- Basque Center on Cognition Brain, and Language, Donostia-San Sebastian 20009, Spain; Stony Brook University, NY 11794-2500, United States; Ikerbasque Foundation, Spain
| |
Collapse
|
42
|
Burton H, Reeder RM, Holden T, Agato A, Firszt JB. Cortical Regions Activated by Spectrally Degraded Speech in Adults With Single Sided Deafness or Bilateral Normal Hearing. Front Neurosci 2021; 15:618326. [PMID: 33897343 PMCID: PMC8058229 DOI: 10.3389/fnins.2021.618326] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Accepted: 03/04/2021] [Indexed: 11/13/2022] Open
Abstract
Those with profound sensorineural hearing loss from single sided deafness (SSD) generally experience greater cognitive effort and fatigue in adverse sound environments. We studied cases with right ear, SSD compared to normal hearing (NH) individuals. SSD cases were significantly less correct in naming last words in spectrally degraded 8- and 16-band vocoded sentences, despite high semantic predictability. Group differences were not significant for less intelligible 4-band sentences, irrespective of predictability. SSD also had diminished BOLD percent signal changes to these same sentences in left hemisphere (LH) cortical regions of early auditory, association auditory, inferior frontal, premotor, inferior parietal, dorsolateral prefrontal, posterior cingulate, temporal-parietal-occipital junction, and posterior opercular. Cortical regions with lower amplitude responses in SSD than NH were mostly components of a LH language network, previously noted as concerned with speech recognition. Recorded BOLD signal magnitudes were averages from all vertices within predefined parcels from these cortex regions. Parcels from different regions in SSD showed significantly larger signal magnitudes to sentences of greater intelligibility (e.g., 8- or 16- vs. 4-band) in all except early auditory and posterior cingulate cortex. Significantly lower response magnitudes occurred in SSD than NH in regions prior studies found responsible for phonetics and phonology of speech, cognitive extraction of meaning, controlled retrieval of word meaning, and semantics. The findings suggested reduced activation of a LH fronto-temporo-parietal network in SSD contributed to difficulty processing speech for word meaning and sentence semantics. Effortful listening experienced by SSD might reflect diminished activation to degraded speech in the affected LH language network parcels. SSD showed no compensatory activity in matched right hemisphere parcels.
Collapse
Affiliation(s)
- Harold Burton
- Department of Neuroscience, Washington University School of Medicine, Saint Louis, MO, United States
| | - Ruth M Reeder
- Department of Otolaryngology-Head and Neck Surgery, Washington University School of Medicine, Saint Louis, MO, United States
| | - Tim Holden
- Department of Otolaryngology-Head and Neck Surgery, Washington University School of Medicine, Saint Louis, MO, United States
| | - Alvin Agato
- Department of Neuroscience, Washington University School of Medicine, Saint Louis, MO, United States
| | - Jill B Firszt
- Department of Otolaryngology-Head and Neck Surgery, Washington University School of Medicine, Saint Louis, MO, United States
| |
Collapse
|
43
|
Kuk F, Slugocki C, Korhonen P. Using the Repeat-Recall Test to Examine Factors Affecting Context Use. J Am Acad Audiol 2021; 31:771-780. [PMID: 33588513 DOI: 10.1055/s-0040-1719136] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
BACKGROUND The effect of context on speech processing has been studied using different speech materials and response criteria. The Repeat-Recall Test (RRT) evaluates listener performance using high context (HC) and low context (LC) sentences; this may offer another platform for studying context use (CU). OBJECTIVE This article aims to evaluate if the RRT may be used to study how different signal-to-noise ratios (SNRs), hearing aid technologies (directional microphone and noise reduction), and listener working memory capacities (WMCs) interact to affect CU on the different measures of the RRT. DESIGN Double-blind, within-subject repeated measures design. STUDY SAMPLE Nineteen listeners with a mild-to-moderately severe hearing loss. DATA COLLECTION The RRT was administered with participants wearing the study hearing aids under two microphone (omnidirectional vs. directional) by two noise reduction (on vs. off) conditions. Speech was presented from 0 degree at 75 dB sound pressure level and a continuous speech-shaped noise from 180 degrees at SNRs of 0, 5, 10, and 15 dB. The order of SNR and hearing aid conditions was counterbalanced across listeners. Each test condition was completed twice in two 2-hour sessions separated by 1 month. RESULTS CU was calculated as the difference between HC and LC sentence scores for each outcome measure (i.e., repeat, recall, listening effort, and tolerable time). For all outcome measures, repeated measures analyses of variance revealed that CU was significantly affected by the SNR of the test conditions. For repeat, recall, and listening effort measures, these effects were qualified by significant two-way interactions between SNR and microphone mode. In addition, the WMC group significantly affected CU during recall and rating of listening effort, the latter of which was qualified by an interaction between the WMC group and SNR. Listener WMC affected CU on estimates of tolerable time as qualified by significant two-way interactions between SNR and microphone mode. CONCLUSION The study supports use of the RRT as a tool for measuring how listeners use sentence context to aid in speech processing. The degree to which context influenced scores on each outcome measure of the RRT was found to depend on complex interactions between the SNR of the listening environment, hearing aid features, and the WMC of the listeners.
Collapse
Affiliation(s)
- Francis Kuk
- WS Audiology, Widex Office of Research in Clinical Amplification (ORCA-USA), Lisle, Illinois
| | - Christopher Slugocki
- WS Audiology, Widex Office of Research in Clinical Amplification (ORCA-USA), Lisle, Illinois
| | - Petri Korhonen
- WS Audiology, Widex Office of Research in Clinical Amplification (ORCA-USA), Lisle, Illinois
| |
Collapse
|
44
|
The Neural Bases of Tinnitus: Lessons from Deafness and Cochlear Implants. J Neurosci 2021; 40:7190-7202. [PMID: 32938634 DOI: 10.1523/jneurosci.1314-19.2020] [Citation(s) in RCA: 51] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2019] [Revised: 08/05/2020] [Accepted: 08/08/2020] [Indexed: 02/06/2023] Open
Abstract
Subjective tinnitus is the conscious perception of sound in the absence of any acoustic source. The literature suggests various tinnitus mechanisms, most of which invoke changes in spontaneous firing rates of central auditory neurons resulting from modification of neural gain. Here, we present an alternative model based on evidence that tinnitus is: (1) rare in people who are congenitally deaf, (2) common in people with acquired deafness, and (3) potentially suppressed by active cochlear implants used for hearing restoration. We propose that tinnitus can only develop after fast auditory fiber activity has stimulated the synapse formation between fast-spiking parvalbumin positive (PV+) interneurons and projecting neurons in the ascending auditory path and coactivated frontostriatal networks after hearing onset. Thereafter, fast auditory fiber activity promotes feedforward and feedback inhibition mediated by PV+ interneuron activity in auditory-specific circuits. This inhibitory network enables enhanced stimulus resolution, attention-driven contrast improvement, and augmentation of auditory responses in central auditory pathways (neural gain) after damage of slow auditory fibers. When fast auditory fiber activity is lost, tonic PV+ interneuron activity is diminished, resulting in the prolonged response latencies, sudden hyperexcitability, enhanced cortical synchrony, elevated spontaneous γ oscillations, and impaired attention/stress-control that have been described in previous tinnitus models. Moreover, because fast processing is gained through sensory experience, tinnitus would not exist in congenital deafness. Electrical cochlear stimulation may have the potential to reestablish tonic inhibitory networks and thus suppress tinnitus. The proposed framework unites many ideas of tinnitus pathophysiology and may catalyze cooperative efforts to develop tinnitus therapies.
Collapse
|
45
|
Abstract
In everyday language processing, sentence context affects how readers and listeners process upcoming words. In experimental situations, it can be useful to identify words that are predicted to greater or lesser degrees by the preceding context. Here we report completion norms for 3085 English sentences, collected online using a written cloze procedure in which participants were asked to provide their best guess for the word completing a sentence. Sentences varied between eight and ten words in length. At least 100 unique participants contributed to each sentence. All responses were reviewed by human raters to mitigate the influence of mis-spellings and typographical errors. The responses provide a range of predictability values for 13,438 unique target words, 6790 of which appear in more than one sentence context. We also provide entropy values based on the relative predictability of multiple responses. A searchable set of norms is available at http://sentencenorms.net. Finally, we provide the code used to collate and organize the responses to facilitate additional analyses and future research projects.
Collapse
|
46
|
Adaptation to mis-pronounced speech: evidence for a prefrontal-cortex repair mechanism. Sci Rep 2021; 11:97. [PMID: 33420193 PMCID: PMC7794353 DOI: 10.1038/s41598-020-79640-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2020] [Accepted: 11/23/2020] [Indexed: 11/30/2022] Open
Abstract
Speech is a complex and ambiguous acoustic signal that varies significantly within and across speakers. Despite the processing challenge that such variability poses, humans adapt to systematic variations in pronunciation rapidly. The goal of this study is to uncover the neurobiological bases of the attunement process that enables such fluent comprehension. Twenty-four native English participants listened to words spoken by a “canonical” American speaker and two non-canonical speakers, and performed a word-picture matching task, while magnetoencephalography was recorded. Non-canonical speech was created by including systematic phonological substitutions within the word (e.g. [s] → [sh]). Activity in the auditory cortex (superior temporal gyrus) was greater in response to substituted phonemes, and, critically, this was not attenuated by exposure. By contrast, prefrontal regions showed an interaction between the presence of a substitution and the amount of exposure: activity decreased for canonical speech over time, whereas responses to non-canonical speech remained consistently elevated. Grainger causality analyses further revealed that prefrontal responses serve to modulate activity in auditory regions, suggesting the recruitment of top-down processing to decode non-canonical pronunciations. In sum, our results suggest that the behavioural deficit in processing mispronounced phonemes may be due to a disruption to the typical exchange of information between the prefrontal and auditory cortices as observed for canonical speech.
Collapse
|
47
|
Spontaneous brain activity underlying auditory hallucinations in the hearing-impaired. Cortex 2021; 136:1-13. [PMID: 33450598 DOI: 10.1016/j.cortex.2020.12.005] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2020] [Revised: 11/09/2020] [Accepted: 12/05/2020] [Indexed: 11/22/2022]
Abstract
Auditory hallucinations, the perception of a sound without a corresponding source, are common in people with hearing impairment. Two forms can be distinguished: simple (i.e., tinnitus) and complex hallucinations (speech and music). Little is known about the precise mechanisms underlying these types of hallucinations. Here we tested the assumption that spontaneous activity in the auditory pathways, following deafferentation, underlies these hallucinations and is related to their phenomenology. By extracting (fractional) Amplitude of Low Frequency Fluctuation [(f)ALFF] scores from resting state fMRI of 18 hearing impaired patients with complex hallucinations (voices or music), 18 hearing impaired patients with simple hallucinations (tinnitus or murmuring), and 20 controls with normal hearing, we investigated differences in spontaneous brain activity between these groups. Spontaneous activity in the anterior and posterior cingulate cortex of hearing-impaired groups was significantly higher than in the controls. The group with complex hallucinations showed elevated activity in the bilateral temporal cortex including Wernicke's area, while spontaneous activity of the group with simple hallucinations was mainly located in the cerebellum. These results suggest a decrease in error monitoring in both hearing-impaired groups. Spontaneous activity of language-related areas only in complex hallucinations suggests that the manifestation of the spontaneous activity represents the phenomenology of the hallucination. The link between cerebellar activity and simple hallucinations, such as tinnitus, is new and may have consequences for treatment.
Collapse
|
48
|
Kajiura M, Jeong H, Kawata NYS, Yu S, Kinoshita T, Kawashima R, Sugiura M. Brain activity predicts future learning success in intensive second language listening training. BRAIN AND LANGUAGE 2021; 212:104839. [PMID: 33271393 DOI: 10.1016/j.bandl.2020.104839] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2019] [Revised: 06/03/2020] [Accepted: 07/14/2020] [Indexed: 06/12/2023]
Abstract
This study explores neural mechanisms underlying how prior knowledge gained from pre-listening transcript reading helps comprehend fast-rate speech in a second language (L2) and applies to L2 learning. Top-down predictive processing by prior knowledge may play an important role in L2 speech comprehension and improving listening skill. By manipulating the pre-listening transcript effect (pre-listening transcript reading [TR] vs. no transcript reading [NTR]) and type of languages (first language (L1) vs. L2), we measured brain activity in L2 learners, who performed fast-rate listening comprehension tasks during functional magnetic resonance imaging. Thereafter, we examined whether TR_L2-specific brain activity can predict individual learning success after an intensive listening training. The left angular and superior temporal gyri were key areas responsible for integrating prior knowledge to sensory input. Activity in these areas correlated significantly with gain scores on subsequent training, indicating that brain activity related to prior knowledge-sensory input integration predicts future learning success.
Collapse
Affiliation(s)
- Mayumi Kajiura
- Division of Foreign Language Education, Aichi Shukutoku University, Nagoya, Japan.
| | - Hyeonjeong Jeong
- Graduate School of International Cultural Studies, Tohoku University, Sendai, Japan; Institute of Development, Aging and Cancer, Tohoku University, Sendai, Japan.
| | - Natasha Y S Kawata
- Institute of Development, Aging and Cancer, Tohoku University, Sendai, Japan
| | - Shaoyun Yu
- Graduate School of Humanities, Nagoya University, Nagoya, Japan
| | - Toru Kinoshita
- Graduate School of Humanities, Nagoya University, Nagoya, Japan
| | - Ryuta Kawashima
- Institute of Development, Aging and Cancer, Tohoku University, Sendai, Japan
| | - Motoaki Sugiura
- Institute of Development, Aging and Cancer, Tohoku University, Sendai, Japan; International Research Institute for Disaster Science, Tohoku University, Sendai, Japan
| |
Collapse
|
49
|
Swanborough H, Staib M, Frühholz S. Neurocognitive dynamics of near-threshold voice signal detection and affective voice evaluation. SCIENCE ADVANCES 2020; 6:6/50/eabb3884. [PMID: 33310844 PMCID: PMC7732184 DOI: 10.1126/sciadv.abb3884] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/20/2020] [Accepted: 10/29/2020] [Indexed: 05/10/2023]
Abstract
Communication and voice signal detection in noisy environments are universal tasks for many species. The fundamental problem of detecting voice signals in noise (VIN) is underinvestigated especially in its temporal dynamic properties. We investigated VIN as a dynamic signal-to-noise ratio (SNR) problem to determine the neurocognitive dynamics of subthreshold evidence accrual and near-threshold voice signal detection. Experiment 1 showed that dynamic VIN, including a varying SNR and subthreshold sensory evidence accrual, is superior to similar conditions with nondynamic SNRs or with acoustically matched sounds. Furthermore, voice signals with affective meaning have a detection advantage during VIN. Experiment 2 demonstrated that VIN is driven by an effective neural integration in an auditory cortical-limbic network at and beyond the near-threshold detection point, which is preceded by activity in subcortical auditory nuclei. This demonstrates the superior recognition advantage of communication signals in dynamic noise contexts, especially when carrying socio-affective meaning.
Collapse
Affiliation(s)
- Huw Swanborough
- Cognitive and Affective Neuroscience Unit, Department of Psychology, University of Zurich, Zurich, Switzerland.
- Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Matthias Staib
- Cognitive and Affective Neuroscience Unit, Department of Psychology, University of Zurich, Zurich, Switzerland
- Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Sascha Frühholz
- Cognitive and Affective Neuroscience Unit, Department of Psychology, University of Zurich, Zurich, Switzerland.
- Neuroscience Center Zurich, University of Zurich and ETH Zurich, Zurich, Switzerland
- Department of Psychology, University of Oslo, Oslo, Norway
| |
Collapse
|
50
|
Chan TMV, Alain C. Brain indices associated with semantic cues prior to and after a word in noise. Brain Res 2020; 1751:147206. [PMID: 33189693 DOI: 10.1016/j.brainres.2020.147206] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2020] [Revised: 11/01/2020] [Accepted: 11/09/2020] [Indexed: 10/23/2022]
Abstract
It is well established that identification of words in noise improves when it is preceded by a semantically related word, but comparatively little is known about the effect of subsequent context in guiding word in noise identification. We build on the findings of a previous behavioural study (Chan & Alain, 2019) by measuring neuro-electric brain activity while manipulating the semantic content of a cue that either preceded or followed a word in noise. Participants were more accurate in identifying the word in noise when it was preceded or followed by a cue that was semantically related. This gain in accuracy coincided with a late positive component, which was time-locked to the word in noise when preceded by a cue and time-locked to the cue when it followed the word in noise. Distributed source analyses of this positive component revealed different patterns in source activity between the two temporal conditions. The effects of relatedness also generated an event-related potential modulation around 400 ms (N400) that was present at cue presentation when it followed the word in noise, but not for the word in noise when preceded by the cue, consistent with findings regarding its sensitivity to signal degradation. Exploratory analyses examined a subset of data based on participants' subjective perceived clarity, which revealed a posterior deflection over the left hemisphere that showed a relatedness effect. We discuss these findings in light of research on prediction as well as a reflective attention framework.
Collapse
Affiliation(s)
- T M Vanessa Chan
- Department of Psychology, University of Toronto, Sidney Smith Building, 100 St. George St., Toronto, Ontario M5S 3G3, Canada; Rotman Research Institute, Baycrest, 3560 Bathurst Street, Toronto, Ontario M6A 2E1, Canada
| | - Claude Alain
- Department of Psychology, University of Toronto, Sidney Smith Building, 100 St. George St., Toronto, Ontario M5S 3G3, Canada; Rotman Research Institute, Baycrest, 3560 Bathurst Street, Toronto, Ontario M6A 2E1, Canada; Institute of Medical Sciences, University of Toronto, Toronto, Ontario, Canada; Faculty of Music, University of Toronto, Toronto, Ontario, Canada.
| |
Collapse
|