1
|
Farrar R, Ashjaei S, Arjmandi MK. Speech-evoked cortical activities and speech recognition in adult cochlear implant listeners: a review of functional near-infrared spectroscopy studies. Exp Brain Res 2024; 242:2509-2530. [PMID: 39305309 PMCID: PMC11527908 DOI: 10.1007/s00221-024-06921-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2024] [Accepted: 09/04/2024] [Indexed: 11/01/2024]
Abstract
Cochlear implants (CIs) are the most successful neural prostheses, enabling individuals with severe to profound hearing loss to access sounds and understand speech. While CI has demonstrated success, speech perception outcomes vary largely among CI listeners, with significantly reduced performance in noise. This review paper summarizes prior findings on speech-evoked cortical activities in adult CI listeners using functional near-infrared spectroscopy (fNIRS) to understand (a) speech-evoked cortical processing in CI listeners compared to normal-hearing (NH) individuals, (b) the relationship between these activities and behavioral speech recognition scores, (c) the extent to which current fNIRS-measured speech-evoked cortical activities in CI listeners account for their differences in speech perception, and (d) challenges in using fNIRS for CI research. Compared to NH listeners, CI listeners had diminished speech-evoked activation in the middle temporal gyrus (MTG) and in the superior temporal gyrus (STG), except one study reporting an opposite pattern for STG. NH listeners exhibited higher inferior frontal gyrus (IFG) activity when listening to CI-simulated speech compared to natural speech. Among CI listeners, higher speech recognition scores correlated with lower speech-evoked activation in the STG, higher activation in the left IFG and left fusiform gyrus, with mixed findings in the MTG. fNIRS shows promise for enhancing our understanding of cortical processing of speech in CI listeners, though findings are mixed. Challenges include test-retest reliability, managing noise, replicating natural conditions, optimizing montage design, and standardizing methods to establish a strong predictive relationship between fNIRS-based cortical activities and speech perception in CI listeners.
Collapse
Affiliation(s)
- Reed Farrar
- Department of Psychology, University of South Carolina, 1512 Pendleton Street, Columbia, SC, 29208, USA
| | - Samin Ashjaei
- Department of Communication Sciences and Disorders, University of South Carolina, 1705 College Street, Columbia, SC, 29208, USA
| | - Meisam K Arjmandi
- Department of Communication Sciences and Disorders, University of South Carolina, 1705 College Street, Columbia, SC, 29208, USA.
- Institute for Mind and Brain, University of South Carolina, Barnwell Street, Columbia, SC, 29208, USA.
| |
Collapse
|
2
|
Jahn KN, Wiegand-Shahani BM, Moturi V, Kashiwagura ST, Doak KR. Cochlear-implant simulated spectral degradation attenuates emotional responses to environmental sounds. Int J Audiol 2024:1-7. [PMID: 39146030 DOI: 10.1080/14992027.2024.2385552] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2023] [Accepted: 07/22/2024] [Indexed: 08/17/2024]
Abstract
OBJECTIVE Cochlear implants (CI) provide users with a spectrally degraded acoustic signal that could impact their auditory emotional experiences. This study evaluated the effects of CI-simulated spectral degradation on emotional valence and arousal elicited by environmental sounds. DESIGN Thirty emotionally evocative sounds were filtered through a noise-band vocoder. Participants rated the perceived valence and arousal elicited by each of the full-spectrum and vocoded stimuli. These ratings were compared across acoustic conditions (full-spectrum, vocoded) and as a function of stimulus type (unpleasant, neutral, pleasant). STUDY SAMPLE Twenty-five young adults (age 19 to 34 years) with normal hearing. RESULTS Emotional responses were less extreme for spectrally degraded (i.e., vocoded) sounds than for full-spectrum sounds. Specifically, spectrally degraded stimuli were perceived as more negative and less arousing than full-spectrum stimuli. CONCLUSION By meticulously replicating CI spectral degradation while controlling for variables that are confounded within CI users, these findings indicate that CI spectral degradation can compress the range of sound-induced emotion independent of hearing loss and other idiosyncratic device- or person-level variables. Future work will characterize emotional reactions to sound in CI users via objective, psychoacoustic, and subjective measures.
Collapse
Affiliation(s)
- Kelly N Jahn
- Department of Speech, Language, and Hearing, The University of Texas at Dallas, Richardson, TX, USA
- Callier Center for Communication Disorders, The University of Texas at Dallas, Dallas, TX, USA
| | - Braden M Wiegand-Shahani
- Department of Speech, Language, and Hearing, The University of Texas at Dallas, Richardson, TX, USA
- Callier Center for Communication Disorders, The University of Texas at Dallas, Dallas, TX, USA
| | - Vaishnavi Moturi
- Department of Speech, Language, and Hearing, The University of Texas at Dallas, Richardson, TX, USA
| | - Sean Takamoto Kashiwagura
- Department of Speech, Language, and Hearing, The University of Texas at Dallas, Richardson, TX, USA
- Callier Center for Communication Disorders, The University of Texas at Dallas, Dallas, TX, USA
| | - Karlee R Doak
- Department of Speech, Language, and Hearing, The University of Texas at Dallas, Richardson, TX, USA
- Callier Center for Communication Disorders, The University of Texas at Dallas, Dallas, TX, USA
| |
Collapse
|
3
|
Hidalgo C, Zielinski C, Chen S, Roman S, Truy E, Schön D. Similar gaze behaviour during dialogue perception in congenitally deaf children with cochlear Implants and normal hearing children. INTERNATIONAL JOURNAL OF LANGUAGE & COMMUNICATION DISORDERS 2024. [PMID: 39073184 DOI: 10.1111/1460-6984.13094] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Accepted: 07/03/2024] [Indexed: 07/30/2024]
Abstract
BACKGROUND Perceptual and speech production abilities of children with cochlear implants (CIs) are usually tested by word and sentence repetition or naming tests. However, these tests are quite far apart from daily life linguistic contexts. AIM Here, we describe a way of investigating the link between language comprehension and anticipatory verbal behaviour promoting the use of more complex listening situations. METHODS AND PROCEDURE The setup consists in watching the audio-visual dialogue of two actors. Children's gaze switches from one speaker to the other serve as a proxy of their prediction abilities. Moreover, to better understand the basis and the impact of anticipatory behaviour, we also measured children's ability to understand the dialogue content, their speech perception and memory skills as well as their rhythmic skills, that also require temporal predictions. Importantly, we compared children with CI performances with those of an age-matched group of children with normal hearing (NH). OUTCOMES AND RESULTS While children with CI revealed poorer speech perception and verbal working memory abilities than NH children, there was no difference in gaze anticipatory behaviour. Interestingly, in children with CI only, we found a significant correlation between dialogue comprehension, perceptual skills and gaze anticipatory behaviour. CONCLUSION Our results extend to a dialogue context of previous findings showing an absence of predictive deficits in children with CI. The current design seems an interesting avenue to provide an accurate and objective estimate of anticipatory language behaviour in a more ecological linguistic context also with young children. WHAT THIS PAPER ADDS What is already known on the subject Children with cochlear implants seem to have difficulties extracting structure from and learning sequential input patterns, possibly due to signal degradation and auditory deprivation in the first years of life. They also seem to have a reduced use of contextual information and slow language processing among children with hearing loss. What this paper adds to existing knowledge Here we show that when adopting a rather complex linguistic context such as watching a dialogue of two individuals, children with cochlear implants are able to use the speech and language structure to anticipate gaze switches to the upcoming speaker. What are the clinical implications of this work? The present design seems an interesting avenue to provide an accurate and objective estimate of anticipatory behaviour in a more ecological and dynamic linguistic context. Importantly, this measure is implicit and it has been previously used with very young (normal-hearing) children, showing that they spontaneously make anticipatory gaze switches by age two. Thus, this approach may be of interest to refine the speech comprehension assessment at a rather early age after cochlear implantation where explicit behavioural tests are not always reliable and sensitive.
Collapse
Affiliation(s)
- Céline Hidalgo
- Aix Marseille Univ, Inserm, INS, Inst Neurosci Syst, Marseille, France
| | - Christelle Zielinski
- Aix-Marseille Univ, Institute of Language, Communication and the Brain, Marseille, France
| | - Sophie Chen
- Aix Marseille Univ, Inserm, INS, Inst Neurosci Syst, Marseille, France
| | - Stéphane Roman
- Aix Marseille Univ, Inserm, INS, Inst Neurosci Syst, Marseille, France
- Pediatric Otolaryngology Department, La Timone Children's Hospital (APHM), Marseille, France
| | - Eric Truy
- Service d'ORL et de Chirurgie cervico-faciale, Hôpital Edouard Herriot, CHU, LYON, France
- Inserm U1028, Lyon Neuroscience Research Center, Equipe IMPACT, Lyon, France
- CNRS UMR5292, Lyon Neuroscience Research Center, Equipe IMPACT, Lyon, France
- University Lyon 1, Lyon, France
| | - Daniele Schön
- Aix Marseille Univ, Inserm, INS, Inst Neurosci Syst, Marseille, France
- Aix-Marseille Univ, Institute of Language, Communication and the Brain, Marseille, France
| |
Collapse
|
4
|
Arjmandi MK, Behroozmand R. On the interplay between speech perception and production: insights from research and theories. Front Neurosci 2024; 18:1347614. [PMID: 38332858 PMCID: PMC10850291 DOI: 10.3389/fnins.2024.1347614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Accepted: 01/08/2024] [Indexed: 02/10/2024] Open
Abstract
The study of spoken communication has long been entrenched in a debate surrounding the interdependence of speech production and perception. This mini review summarizes findings from prior studies to elucidate the reciprocal relationships between speech production and perception. We also discuss key theoretical perspectives relevant to speech perception-production loop, including hyper-articulation and hypo-articulation (H&H) theory, speech motor theory, direct realism theory, articulatory phonology, the Directions into Velocities of Articulators (DIVA) and Gradient Order DIVA (GODIVA) models, and predictive coding. Building on prior findings, we propose a revised auditory-motor integration model of speech and provide insights for future research in speech perception and production, focusing on the effects of impaired peripheral auditory systems.
Collapse
Affiliation(s)
- Meisam K. Arjmandi
- Translational Auditory Neuroscience Lab, Department of Communication Sciences and Disorders, Arnold School of Public Health, University of South Carolina, Columbia, SC, United States
| | - Roozbeh Behroozmand
- Speech Neuroscience Lab, Department of Speech, Language, and Hearing, Callier Center for Communication Disorders, School of Behavioral and Brain Sciences, The University of Texas at Dallas, Richardson, TX, United States
| |
Collapse
|
5
|
Skidmore J, Oleson JJ, Yuan Y, He S. The Relationship Between Cochlear Implant Speech Perception Outcomes and Electrophysiological Measures of the Electrically Evoked Compound Action Potential. Ear Hear 2023; 44:1485-1497. [PMID: 37194125 DOI: 10.1097/aud.0000000000001389] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/18/2023]
Abstract
OBJECTIVE This study assessed the relationship between electrophysiological measures of the electrically evoked compound action potential (eCAP) and speech perception scores measured in quiet and in noise in postlingually deafened adult cochlear implant (CI) users. It tested the hypothesis that how well the auditory nerve (AN) responds to electrical stimulation is important for speech perception with a CI in challenging listening conditions. DESIGN Study participants included 24 postlingually deafened adult CI users. All participants used Cochlear Nucleus CIs in their test ears. In each participant, eCAPs were measured at multiple electrode locations in response to single-pulse, paired-pulse, and pulse-train stimuli. Independent variables included six metrics calculated from the eCAP recordings: the electrode-neuron interface (ENI) index, the neural adaptation (NA) ratio, NA speed, the adaptation recovery (AR) ratio, AR speed, and the amplitude modulation (AM) ratio. The ENI index quantified the effectiveness of the CI electrodes in stimulating the targeted AN fibers. The NA ratio indicated the amount of NA at the AN caused by a train of constant-amplitude pulses. NA speed was defined as the speed/rate of NA. The AR ratio estimated the amount of recovery from NA at a fixed time point after the cessation of pulse-train stimulation. AR speed referred to the speed of recovery from NA caused by previous pulse-train stimulation. The AM ratio provided a measure of AN sensitivity to AM cues. Participants' speech perception scores were measured using Consonant-Nucleus-Consonant (CNC) word lists and AzBio sentences presented in quiet, as well as in noise at signal-to-noise ratios (SNRs) of +10 and +5 dB. Predictive models were created for each speech measure to identify eCAP metrics with meaningful predictive power. RESULTS The ENI index and AR speed individually explained at least 10% of the variance in most of the speech perception scores measured in this study, while the NA ratio, NA speed, the AR ratio, and the AM ratio did not. The ENI index was identified as the only eCAP metric that had unique predictive power for each of the speech test results. The amount of variance in speech perception scores (both CNC words and AzBio sentences) explained by the eCAP metrics increased with increased difficulty under the listening condition. Over half of the variance in speech perception scores measured in +5 dB SNR noise (both CNC words and AzBio sentences) was explained by a model with only three eCAP metrics: the ENI index, NA speed, and AR speed. CONCLUSIONS Of the six electrophysiological measures assessed in this study, the ENI index is the most informative predictor for speech perception performance in CI users. In agreement with the tested hypothesis, the response characteristics of the AN to electrical stimulation are more important for speech perception with a CI in noise than they are in quiet.
Collapse
Affiliation(s)
- Jeffrey Skidmore
- Department of Otolaryngology-Head and Neck Surgery, The Ohio State University, Columbus, Ohio, USA
| | - Jacob J Oleson
- Department of Biostatistics, University of Iowa, Iowa City, Iowa, USA
| | - Yi Yuan
- Department of Otolaryngology-Head and Neck Surgery, The Ohio State University, Columbus, Ohio, USA
| | - Shuman He
- Department of Otolaryngology-Head and Neck Surgery, The Ohio State University, Columbus, Ohio, USA
- Department of Audiology, Nationwide Children's Hospital, Columbus, Ohio, USA
| |
Collapse
|
6
|
de la Cruz-Pavía I, Eloy C, Perrineau-Hecklé P, Nazzi T, Cabrera L. Consonant bias in adult lexical processing under acoustically degraded listening conditions. JASA EXPRESS LETTERS 2023; 3:2892558. [PMID: 37220232 DOI: 10.1121/10.0019576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Accepted: 05/05/2023] [Indexed: 05/25/2023]
Abstract
Consonants facilitate lexical processing across many languages, including French. This study investigates whether acoustic degradation affects this phonological bias in an auditory lexical decision task. French words were processed using an eight-band vocoder, degrading their frequency modulations (FM) while preserving original amplitude modulations (AM). Adult French natives were presented with these French words, preceded by similarly processed pseudoword primes sharing their vowels, consonants, or neither. Results reveal a consonant bias in the listeners' accuracy and response times, despite the reduced spectral and FM information. These degraded conditions resemble current cochlear-implant processors, and attest to the robustness of this phonological bias.
Collapse
Affiliation(s)
- Irene de la Cruz-Pavía
- Department of Linguistics and Basque Studies, Universidad del País Vasco/Euskal Herriko Unibertsitatea, Vitoria-Gasteiz 01006, Spain
| | - Coraline Eloy
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Paula Perrineau-Hecklé
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Thierry Nazzi
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Laurianne Cabrera
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| |
Collapse
|