1
|
Marx M, Laborde ML, Algans C, Tartayre M, James CJ. Barriers to Early Progress in Adult Cochlear Implant Outcomes. Ear Hear 2024:00003446-990000000-00329. [PMID: 39129126 DOI: 10.1097/aud.0000000000001559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/13/2024]
Abstract
OBJECTIVES Adult cochlear implant (CI) recipients obtain varying levels of speech perception from their device. Adult CI users adapt quickly to their CI if they have no peripheral "bottom-up" or neurocognitive "top-down" limiting factors. Our objective here was to understand the influence of limiting factors on the progression of sentence understanding in quiet and in noise, initially and over time. We hypothesized that the presence of limiting factors, detected using a short test battery, would predictably influence sentence recognition with practical consequences. We aimed to validate the test battery by comparing the presence of limiting factors and the success criteria of >90% sentence understanding in quiet 1 month after activation. DESIGN The study was a single-clinic, cross-sectional, retrospective design incorporating 32 adult unilateral Nucleus CI users aged 27 to 90 years (mean = 70, SD = 13.5). Postoperative outcome was assessed through sentence recognition scores in quiet and in varying signal to noise ratios at 1 day, 1 to 2 months, and up to 2 years. Our clinic's standard test battery comprises physiological and neurocognitive measures. Physiological measures included electrically evoked compound action potentials for recovery function, spread of excitation, and polarity effect. To evaluate general cognitive function, inhibition, and phonological awareness, the Montreal Cognitive Assessment screening test, the Stroop Color-Word Test, and tests 3 and 4 of the French Assessment of Reading Skills in Adults over 16 years of age, respectively were performed. Physiological scores were considered abnormal, and therefore limiting, when total neural recovery periods and polarity effects, for both apical and basal electrode positions, were >1.65 SDs from the population mean. A spread of excitation of >6 electrode units was also considered limiting. For the neurocognitive tests, scores poorer than 1.65 SDs from published normal population means were considered limiting. RESULTS At 1 month, 13 out of 32 CI users scored ≥90% sentence recognition in quiet with no significant dependence on age. Subjects with no limiting peripheral or neurocognitive factors were 8.5 times more likely to achieve ≥90% score in quiet at 1 month after CI switch-on (p = 0.010). In our sample, we detected 4 out of 32 cases with peripheral limiting factors that related to neural health or poor electrode-neural interface at both apical and basal positions. In contrast, neurocognitive limiting factors were identified in 14 out of 32 subjects. Early sentence recognition scores were predictive of long-term sentence recognition thresholds in noise such that limiting factors appeared to be of continuous influence. CONCLUSIONS Both peripheral and neurocognitive processing factors affect early sentence recognition after CI activation. Peripheral limiting factors may have been detected less often than neurocognitive limiting factors because they were defined using sample-based criteria versus normal population-based criteria. Early performance was generally predictive of long-term performance. Understanding the measurable covariables that limit CI performance may inform follow-up and improve counseling. A score of ≥90% for sentence recognition in quiet at 1 month may be used to define successful progress; whereas, lower scores indicate the need for diagnostic testing and ongoing rehabilitation. Our findings suggest that sentence test scores as early as 1 day after activation can provide vital information for the new CI user and indicate the need for rehabilitation follow-up.
Collapse
Affiliation(s)
- Mathieu Marx
- Service Oto Rhino Laryngologie Hôpital Riquet, Toulouse, France
| | | | - Carol Algans
- Service Oto Rhino Laryngologie Hôpital Riquet, Toulouse, France
| | | | | |
Collapse
|
2
|
Dillon MT, Buss E, Johnson AD, Canfarotta MW, O’Connell BP. Comparison of Two Place-Based Mapping Procedures on Masked Sentence Recognition as a Function of Electrode Array Angular Insertion Depth and Presence of Acoustic Low-Frequency Information: A Simulation Study. Audiol Neurootol 2023; 28:478-487. [PMID: 37482054 PMCID: PMC10948008 DOI: 10.1159/000531262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2022] [Accepted: 05/23/2023] [Indexed: 07/25/2023] Open
Abstract
INTRODUCTION Cochlear implant (CI) and electric-acoustic stimulation (EAS) users may experience better performance with maps that align the electric filter frequencies to the cochlear place frequencies, known as place-based maps, than with maps that present spectrally shifted information. Individual place-based mapping procedures differ in the frequency content that is aligned to cochlear tonotopicity versus discarded or spectrally shifted. The performance benefit with different place-based maps may vary due to individual differences in angular insertion depth (AID) of the electrode array and whether functional acoustic low-frequency information is available in the implanted ear. The present study compared masked speech recognition with two types of place-based maps as a function of AID and presence of acoustic low-frequency information. METHODS Sixty adults with normal hearing listened acutely to CI or EAS simulations of two types of place-based maps for one of three cases of electrode arrays at shallow AIDs. The strict place-based (Strict-PB) map aligned the low- and mid-frequency information to cochlear tonotopicity and discarded information below the frequency associated with the most apical electrode contact. The alternative place-based map (LFshift-PB) aligned the mid-frequency information to cochlear tonotopicity and provided more of the speech spectrum by compressing low-frequency information on the apical electrode contacts (i.e., <1 kHz). Three actual cases of a 12-channel, 24-mm electrode array were simulated by assigning the carrier frequency for an individual channel as the cochlear place frequency of the associated electrode contact. The AID and cochlear place frequency for the most apical electrode contact were 460° and 498 Hz for case 1, 389° and 728 Hz for case 2, and 335° and 987 Hz for case 3, respectively. RESULTS Generally, better performance was observed with the Strict-PB maps for cases 1 and 2, where mismatches were 2-4 octaves for the most apical channel with the LFshift-PB map. Similar performance was observed between maps for case 3. For the CI simulations, performance with the Strict-PB map declined with decreases in AID, while performance with the LFshift-PB map remained stable across cases. For the EAS simulations, performance with the Strict-PB map remained stable across cases, while performance with the LFshift-PB map improved with decreases in AID. CONCLUSIONS Listeners demonstrated differences with the Strict-PB versus LFshift-PB maps as a function of AID and whether acoustic low-frequency information was available (CI vs. EAS). These data support the use of the Strict-PB mapping procedure for AIDs ≥335°, though further study including time for acclimatization in CI and EAS users is warranted.
Collapse
Affiliation(s)
- Margaret T. Dillon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Emily Buss
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Alec D. Johnson
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Michael W. Canfarotta
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Brendan P. O’Connell
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
- Charlotte Eye Ear Nose & Throat Associates, P.A., Charlotte, NC, USA
| |
Collapse
|
3
|
Dillon MT, Canfarotta MW, Buss E, Rooth MA, Richter ME, Overton AB, Roth NE, Dillon SM, Raymond JH, Young A, Pearson AC, Davis AG, Dedmon MM, Brown KD, O'Connell BP. Influence of Electric Frequency-to-Place Mismatches on the Early Speech Recognition Outcomes for Electric-Acoustic Stimulation Users. Am J Audiol 2023; 32:251-260. [PMID: 36800505 PMCID: PMC10166189 DOI: 10.1044/2022_aja-21-00254] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2021] [Revised: 05/21/2022] [Accepted: 11/28/2022] [Indexed: 02/19/2023] Open
Abstract
PURPOSE Cochlear implant (CI) recipients with hearing preservation experience significant improvements in speech recognition with electric-acoustic stimulation (EAS) as compared to with a CI alone, although outcomes across EAS users vary. The individual differences in performance may be due in part to default mapping procedures, which result in electric frequency-to-place mismatches for the majority of EAS users. This study assessed the influence of electric mismatches on the early speech recognition for EAS users. METHOD Twenty-one participants were randomized at EAS activation to listen exclusively with a default or place-based map. For both groups, the unaided thresholds determined the acoustic cutoff frequency (i.e., > 65 dB HL). For default maps, the electric filter frequencies were assigned to avoid spectral gaps in frequency information but created varying magnitudes of mismatches. For place-based maps, the electric filter frequencies were assigned to avoid frequency-to-place mismatches. Recognition of consonant-nucleus-consonant words and vowels was assessed at activation and 1, 3, and 6 months postactivation. RESULTS For participants with default maps, electric mismatch at 1500 Hz ranged from 2 to -12.0 semitones (Mdn = -5 semitones). Poorer performance was observed for those with larger magnitudes of electric mismatch. This effect was observed through 6 months of EAS listening experience. CONCLUSIONS The present sample of EAS users experienced better initial performance when electric mismatches were small or eliminated. These data suggest the utility of methods that reduce electric mismatches, such as place-based mapping procedures. Investigation is ongoing to determine whether these differences persist with long-term EAS use. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.22096523.
Collapse
Affiliation(s)
- Margaret T. Dillon
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, The University of North Carolina at Chapel Hill
| | - Michael W. Canfarotta
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| | - Emily Buss
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| | - Meredith A. Rooth
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| | - Margaret E. Richter
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, The University of North Carolina at Chapel Hill
| | | | | | | | | | - Allison Young
- Department of Audiology, UNC Health, Chapel Hill, NC
| | | | - Amanda G. Davis
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, The University of North Carolina at Chapel Hill
| | - Matthew M. Dedmon
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| | - Kevin D. Brown
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| | - Brendan P. O'Connell
- Department of Otolaryngology/Head & Neck Surgery, School of Medicine, The University of North Carolina at Chapel Hill
| |
Collapse
|
4
|
Dillon MT, O'Connell BP, Canfarotta MW, Buss E, Hopfinger J. Effect of Place-Based Versus Default Mapping Procedures on Masked Speech Recognition: Simulations of Cochlear Implant Alone and Electric-Acoustic Stimulation. Am J Audiol 2022; 31:322-337. [PMID: 35394798 DOI: 10.1044/2022_aja-21-00123] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE Cochlear implant (CI) recipients demonstrate variable speech recognition when listening with a CI-alone or electric-acoustic stimulation (EAS) device, which may be due in part to electric frequency-to-place mismatches created by the default mapping procedures. Performance may be improved if the filter frequencies are aligned with the cochlear place frequencies, known as place-based mapping. Performance with default maps versus an experimental place-based map was compared for participants with normal hearing when listening to CI-alone or EAS simulations to observe potential outcomes prior to initiating an investigation with CI recipients. METHOD A noise vocoder simulated CI-alone and EAS devices, mapped with default or place-based procedures. The simulations were based on an actual 24-mm electrode array recipient, whose insertion angles for each electrode contact were used to estimate the respective cochlear place frequency. The default maps used the filter frequencies assigned by the clinical software. The filter frequencies for the place-based maps aligned with the cochlear place frequencies for individual contacts in the low- to mid-frequency cochlear region. For the EAS simulations, low-frequency acoustic information was filtered to simulate aided low-frequency audibility. Performance was evaluated for the AzBio sentences presented in a 10-talker masker at +5 dB signal-to-noise ratio (SNR), +10 dB SNR, and asymptote. RESULTS Performance was better with the place-based maps as compared with the default maps for both CI-alone and EAS simulations. For instance, median performance at +10 dB SNR for the CI-alone simulation was 57% correct for the place-based map and 20% for the default map. For the EAS simulation, those values were 59% and 37% correct. Adding acoustic low-frequency information resulted in a similar benefit for both maps. CONCLUSIONS Reducing frequency-to-place mismatches, such as with the experimental place-based mapping procedure, produces a greater benefit in speech recognition than maximizing bandwidth for CI-alone and EAS simulations. Ongoing work is evaluating the initial and long-term performance benefits in CI-alone and EAS users. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.19529053.
Collapse
Affiliation(s)
- Margaret T. Dillon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, University of North Carolina at Chapel Hill
| | - Brendan P. O'Connell
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Michael W. Canfarotta
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Emily Buss
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Joseph Hopfinger
- Department of Psychology and Neuroscience, University of North Carolina at Chapel Hill
| |
Collapse
|
5
|
Xu K, Willis S, Gopen Q, Fu QJ. Effects of Spectral Resolution and Frequency Mismatch on Speech Understanding and Spatial Release From Masking in Simulated Bilateral Cochlear Implants. Ear Hear 2021; 41:1362-1371. [PMID: 32132377 DOI: 10.1097/aud.0000000000000865] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Due to interaural frequency mismatch, bilateral cochlear-implant (CI) users may be less able to take advantage of binaural cues that normal-hearing (NH) listeners use for spatial hearing, such as interaural time differences and interaural level differences. As such, bilateral CI users have difficulty segregating competing speech even when the target and competing talkers are spatially separated. The goal of this study was to evaluate the effects of spectral resolution, tonotopic mismatch (the frequency mismatch between the acoustic center frequency assigned to CI electrode within an implanted ear relative to the expected spiral ganglion characteristic frequency), and interaural mismatch (differences in the degree of tonotopic mismatch in each ear) on speech understanding and spatial release from masking (SRM) in the presence of competing talkers in NH subjects listening to bilateral vocoder simulations. DESIGN During testing, both target and masker speech were presented in five-word sentences that had the same syntax but were not necessarily meaningful. The sentences were composed of five categories in fixed order (Name, Verb, Number, Color, and Clothes), each of which had 10 items, such that multiple sentences could be generated by randomly selecting a word from each category. Speech reception thresholds (SRTs) for the target sentence presented in competing speech maskers were measured. The target speech was delivered to both ears and the two speech maskers were delivered to (1) both ears (diotic masker), or (2) different ears (dichotic masker: one delivered to the left ear and the other delivered to the right ear). Stimuli included the unprocessed speech and four 16-channel sine-vocoder simulations with different interaural mismatch (0, 1, and 2 mm). SRM was calculated as the difference between the diotic and dichotic listening conditions. RESULTS With unprocessed speech, SRTs were 0.3 and -18.0 dB for the diotic and dichotic maskers, respectively. For the spectrally degraded speech with mild tonotopic mismatch and no interaural mismatch, SRTs were 5.6 and -2.0 dB for the diotic and dichotic maskers, respectively. When the tonotopic mismatch increased in both ears, SRTs worsened to 8.9 and 2.4 dB for the diotic and dichotic maskers, respectively. When the two ears had different tonotopic mismatch (e.g., there was interaural mismatch), the performance drop in SRTs was much larger for the dichotic than for the diotic masker. The largest SRM was observed with unprocessed speech (18.3 dB). With the CI simulations, SRM was significantly reduced to 7.6 dB even with mild tonotopic mismatch but no interaural mismatch; SRM was further reduced with increasing interaural mismatch. CONCLUSIONS The results demonstrate that frequency resolution, tonotopic mismatch, and interaural mismatch have differential effects on speech understanding and SRM in simulation of bilateral CIs. Minimizing interaural mismatch may be critical to optimize binaural benefits and improve CI performance for competing speech, a typical listening environment. SRM (the difference in SRTs between diotic and dichotic maskers) may be a useful clinical tool to assess interaural frequency mismatch in bilateral CI users and to evaluate the benefits of optimization methods that minimize interaural mismatch.
Collapse
Affiliation(s)
- Kevin Xu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, California, USA
| | | | | | | |
Collapse
|
6
|
Individual Variability in Recalibrating to Spectrally Shifted Speech: Implications for Cochlear Implants. Ear Hear 2021; 42:1412-1427. [PMID: 33795617 DOI: 10.1097/aud.0000000000001043] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Cochlear implant (CI) recipients are at a severe disadvantage compared with normal-hearing listeners in distinguishing consonants that differ by place of articulation because the key relevant spectral differences are degraded by the implant. One component of that degradation is the upward shifting of spectral energy that occurs with a shallow insertion depth of a CI. The present study aimed to systematically measure the effects of spectral shifting on word recognition and phoneme categorization by specifically controlling the amount of shifting and using stimuli whose identification specifically depends on perceiving frequency cues. We hypothesized that listeners would be biased toward perceiving phonemes that contain higher-frequency components because of the upward frequency shift and that intelligibility would decrease as spectral shifting increased. DESIGN Normal-hearing listeners (n = 15) heard sine wave-vocoded speech with simulated upward frequency shifts of 0, 2, 4, and 6 mm of cochlear space to simulate shallow CI insertion depth. Stimuli included monosyllabic words and /b/-/d/ and /∫/-/s/ continua that varied systematically by formant frequency transitions or frication noise spectral peaks, respectively. Recalibration to spectral shifting was operationally defined as shifting perceptual acoustic-phonetic mapping commensurate with the spectral shift. In other words, adjusting frequency expectations for both phonemes upward so that there is still a perceptual distinction, rather than hearing all upward-shifted phonemes as the higher-frequency member of the pair. RESULTS For moderate amounts of spectral shifting, group data suggested a general "halfway" recalibration to spectral shifting, but individual data suggested a notably different conclusion: half of the listeners were able to recalibrate fully, while the other halves of the listeners were utterly unable to categorize shifted speech with any reliability. There were no participants who demonstrated a pattern intermediate to these two extremes. Intelligibility of words decreased with greater amounts of spectral shifting, also showing loose clusters of better- and poorer-performing listeners. Phonetic analysis of word errors revealed certain cues were more susceptible to being compromised due to a frequency shift (place and manner of articulation), while voicing was robust to spectral shifting. CONCLUSIONS Shifting the frequency spectrum of speech has systematic effects that are in line with known properties of speech acoustics, but the ensuing difficulties cannot be predicted based on tonotopic mismatch alone. Difficulties are subject to substantial individual differences in the capacity to adjust acoustic-phonetic mapping. These results help to explain why speech recognition in CI listeners cannot be fully predicted by peripheral factors like electrode placement and spectral resolution; even among listeners with functionally equivalent auditory input, there is an additional factor of simply being able or unable to flexibly adjust acoustic-phonetic mapping. This individual variability could motivate precise treatment approaches guided by an individual's relative reliance on wideband frequency representation (even if it is mismatched) or limited frequency coverage whose tonotopy is preserved.
Collapse
|
7
|
Effects of noise on integration of acoustic and electric hearing within and across ears. PLoS One 2020; 15:e0240752. [PMID: 33057396 PMCID: PMC7561114 DOI: 10.1371/journal.pone.0240752] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2020] [Accepted: 10/01/2020] [Indexed: 11/19/2022] Open
Abstract
In bimodal listening, cochlear implant (CI) users combine electric hearing (EH) in one ear and acoustic hearing (AH) in the other ear. In electric-acoustic stimulation (EAS), CI users combine EH and AH in the same ear. In quiet, integration of EH and AH has been shown to be better with EAS, but with greater sensitivity to tonotopic mismatch in EH. The goal of the present study was to evaluate how external noise might affect integration of AH and EH within or across ears. Recognition of monosyllabic words was measured for normal-hearing subjects listening to simulations of unimodal (AH or EH alone), EAS, and bimodal listening in quiet and in speech-shaped steady noise (10 dB, 0 dB signal-to-noise ratio). The input/output frequency range for AH was 0.1–0.6 kHz. EH was simulated using an 8-channel noise vocoder. The output frequency range was 1.2–8.0 kHz to simulate a shallow insertion depth. The input frequency range was either matched (1.2–8.0 kHz) or mismatched (0.6–8.0 kHz) to the output frequency range; the mismatched input range maximized the amount of speech information, while the matched input resulted in some speech information loss. In quiet, tonotopic mismatch differently affected EAS and bimodal performance. In noise, EAS and bimodal performance was similarly affected by tonotopic mismatch. The data suggest that tonotopic mismatch may differently affect integration of EH and AH in quiet and in noise.
Collapse
|
8
|
Lambriks LJG, van Hoof M, Debruyne JA, Janssen M, Chalupper J, van der Heijden KA, Hof JR, Hellingman CA, George ELJ, Devocht EMJ. Evaluating hearing performance with cochlear implants within the same patient using daily randomization and imaging-based fitting - The ELEPHANT study. Trials 2020; 21:564. [PMID: 32576247 PMCID: PMC7310427 DOI: 10.1186/s13063-020-04469-x] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2020] [Accepted: 05/30/2020] [Indexed: 02/08/2023] Open
Abstract
Background Prospective research in the field of cochlear implants is hampered by methodological issues and small sample sizes. The ELEPHANT study presents an alternative clinical trial design with a daily randomized approach evaluating individualized tonotopical fitting of a cochlear implant (CI). Methods A single-blinded, daily-randomized clinical trial will be implemented to evaluate a new imaging-based CI mapping strategy. A minimum of 20 participants will be included from the start of the rehabilitation process with a 1-year follow-up period. Based on a post-operative cone beam CT scan (CBCT), mapping of electrical input will be aligned to natural place-pitch arrangement in the individual cochlea. The CI’s frequency allocation table will be adjusted to match the electrical stimulation of frequencies as closely as possible to corresponding acoustic locations in the cochlea. A randomization scheme will be implemented whereby the participant, blinded to the intervention allocation, crosses over between the experimental and standard fitting program on a daily basis, and thus effectively acts as his own control, followed by a period of free choice between both maps to incorporate patient preference. With this new approach the occurrence of a first-order carryover effect and a limited sample size is addressed. Discussion The experimental fitting strategy is thought to give rise to a steeper learning curve, result in better performance in challenging listening situations, improve sound quality, better complement residual acoustic hearing in the contralateral ear and be preferred by recipients of a CI. Concurrently, the suitability of the novel trial design will be considered in investigating these hypotheses. Trial registration ClinicalTrials.gov: NCT03892941. Registered 27 March 2019.
Collapse
Affiliation(s)
- L J G Lambriks
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands.
| | - M van Hoof
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - J A Debruyne
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - M Janssen
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands.,Department of Methodology and Statistics, School for Public Health and Primary Care (CAPHRI), Maastricht University Medical Center, Maastricht, The Netherlands
| | - J Chalupper
- Advanced Bionics European Research Centre (AB ERC), Hannover, Germany
| | - K A van der Heijden
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - J R Hof
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - C A Hellingman
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - E L J George
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| | - E M J Devocht
- Department of ENT/Audiology, School for Mental Health and Neuroscience (MHeNs), Maastricht University Medical Center, Maastricht, The Netherlands
| |
Collapse
|
9
|
Karoui C, James C, Barone P, Bakhos D, Marx M, Macherey O. Searching for the Sound of a Cochlear Implant: Evaluation of Different Vocoder Parameters by Cochlear Implant Users With Single-Sided Deafness. Trends Hear 2020; 23:2331216519866029. [PMID: 31533581 PMCID: PMC6753516 DOI: 10.1177/2331216519866029] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Cochlear implantation in subjects with single-sided deafness (SSD) offers a unique opportunity to directly compare the percepts evoked by a cochlear implant (CI) with those evoked acoustically. Here, nine SSD-CI users performed a forced-choice task evaluating the similarity of speech processed by their CI with speech processed by several vocoders presented to their healthy ear. In each trial, subjects heard two intervals: their CI followed by a certain vocoder in Interval 1 and their CI followed by a different vocoder in Interval 2. The vocoders differed either (i) in carrier type-(sinusoidal [SINE], bandfiltered noise [NOISE], and pulse-spreading harmonic complex) or (ii) in frequency mismatch between the analysis and synthesis frequency ranges-(no mismatch, and two frequency-mismatched conditions of 2 and 4 equivalent rectangular bandwidths [ERBs]). Subjects had to state in which of the two intervals the CI and vocoder sounds were more similar. Despite a large intersubject variability, the PSHC vocoder was judged significantly more similar to the CI than SINE or NOISE vocoders. Furthermore, the No-mismatch and 2-ERB mismatch vocoders were judged significantly more similar to the CI than the 4-ERB mismatch vocoder. The mismatch data were also interpreted by comparing spiral ganglion characteristic frequencies with electrode contact positions determined from postoperative computed tomography scans. Only one subject demonstrated a pattern of preference consistent with adaptation to the CI sound processor frequency-to-electrode allocation table and two subjects showed possible partial adaptation. Those subjects with adaptation patterns presented overall small and consistent frequency mismatches across their electrode arrays.
Collapse
Affiliation(s)
- Chadlia Karoui
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Cochlear France SAS, Toulouse, France
| | - Chris James
- Cochlear France SAS, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Pascal Barone
- Centre de Recherche Cerveau et Cognition, Toulouse, France
| | - David Bakhos
- Université François-Rabelais de Tours, CHRU de Tours, France.,Ear Nose and Throat department, CHUR de Tours, Tours, France
| | - Mathieu Marx
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Olivier Macherey
- Aix Marseille University, CNRS, Centrale Marseille, LMA, Marseille, France
| |
Collapse
|
10
|
|
11
|
A Smartphone Application for Customized Frequency Table Selection in Cochlear Implants. Otol Neurotol 2018; 38:e253-e261. [PMID: 28806335 DOI: 10.1097/mao.0000000000001409] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
HYPOTHESIS A novel smartphone-based software application can facilitate self-selection of frequency allocation tables (FAT) in postlingually deaf cochlear implant (CI) users. BACKGROUND CIs use FATs to represent the tonotopic organization of a normal cochlea. Current CI fitting methods typically use a standard FAT for all patients regardless of individual differences in cochlear size and electrode location. In postlingually deaf patients, different amounts of mismatch can result between the frequency-place function they experienced when they had normal hearing and the frequency-place function that results from the standard FAT. For some CI users, an alternative FAT may enhance sound quality or speech perception. Currently, no widely available tools exist to aid real-time selection of different FATs. This study aims to develop a new smartphone tool for this purpose and to evaluate speech perception and sound quality measures in a pilot study of CI subjects using this application. METHODS A smartphone application for a widely available mobile platform (iOS) was developed to serve as a preprocessor of auditory input to a clinical CI speech processor and enable interactive real-time selection of FATs. The application's output was validated by measuring electrodograms for various inputs. A pilot study was conducted in six CI subjects. Speech perception was evaluated using word recognition tests. RESULTS All subjects successfully used the portable application with their clinical speech processors to experience different FATs while listening to running speech. The users were all able to select one table that they judged provided the best sound quality. All subjects chose a FAT different from the standard FAT in their everyday clinical processor. Using the smartphone application, the mean consonant-nucleus-consonant score with the default FAT selection was 28.5% (SD 16.8) and 29.5% (SD 16.4) when using a self-selected FAT. CONCLUSION A portable smartphone application enables CI users to self-select frequency allocation tables in real time. Even though the self-selected FATs that were deemed to have better sound quality were only tested acutely (i.e., without long-term experience with them), speech perception scores were not inferior to those obtained with the clinical FATs. This software application may be a valuable tool for improving future methods of CI fitting.
Collapse
|
12
|
Integration of acoustic and electric hearing is better in the same ear than across ears. Sci Rep 2017; 7:12500. [PMID: 28970567 PMCID: PMC5624923 DOI: 10.1038/s41598-017-12298-3] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2017] [Accepted: 09/06/2017] [Indexed: 11/26/2022] Open
Abstract
Advances in cochlear implant (CI) technology allow for acoustic and electric hearing to be combined within the same ear (electric-acoustic stimulation, or EAS) and/or across ears (bimodal listening). Integration efficiency (IE; the ratio between observed and predicted performance for acoustic-electric hearing) can be used to estimate how well acoustic and electric hearing are combined. The goal of this study was to evaluate factors that affect IE in EAS and bimodal listening. Vowel recognition was measured in normal-hearing subjects listening to simulations of unimodal, EAS, and bimodal listening. The input/output frequency range for acoustic hearing was 0.1–0.6 kHz. For CI simulations, the output frequency range was 1.2–8.0 kHz to simulate a shallow insertion depth and the input frequency range was varied to provide increasing amounts of speech information and tonotopic mismatch. Performance was best when acoustic and electric hearing was combined in the same ear. IE was significantly better for EAS than for bimodal listening; IE was sensitive to tonotopic mismatch for EAS, but not for bimodal listening. These simulation results suggest acoustic and electric hearing may be more effectively and efficiently combined within rather than across ears, and that tonotopic mismatch should be minimized to maximize the benefit of acoustic-electric hearing, especially for EAS.
Collapse
|
13
|
Schvartz-Leyzac KC, Zwolan TA, Pfingst BE. Effects of electrode deactivation on speech recognition in multichannel cochlear implant recipients. Cochlear Implants Int 2017; 18:324-334. [PMID: 28793847 DOI: 10.1080/14670100.2017.1359457] [Citation(s) in RCA: 22] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
OBJECTIVES The objective of the current study is to evaluate how speech recognition performance is affected by the number of active electrodes that are turned off in multichannel cochlear implants. Several recent studies have demonstrated positive effects of deactivating stimulation sites based on an objective measure in cochlear implant processing strategies. Previous studies using an analysis of variance have shown that, on average, cochlear implant listeners' performance does not improve beyond eight active electrodes. We hypothesized that using a generalized linear mixed model would allow for better examination of this question. METHODS Seven peri- and post-lingual adult cochlear implant users (eight ears) were tested on speech recognition tasks using experimental MAPs which contained either 8, 12, 16 or 20 active electrodes. Speech recognition tests included CUNY sentences in speech-shaped noise, TIMIT sentences in quiet as well as vowel (CVC) and consonant (CV) stimuli presented in quiet and in signal-to-noise ratios of 0 and +10 dB. RESULTS The speech recognition threshold in noise (dB SNR) significantly worsened by approximately 2 dB on average as the number of active electrodes was decreased from 20 to 8. Likewise, sentence recognition scores in quiet significantly decreased by an average of approximately 12%. DISCUSSION/CONCLUSION Cochlear implant recipients can utilize and benefit from using more than eight spectral channels when listening to complex sentences or sentences in background noise. The results of the current study suggest a conservative approach for turning off stimulation sites is best when using site-selection procedures.
Collapse
Affiliation(s)
- Kara C Schvartz-Leyzac
- a Department of Otolaryngology , Kresge Hearing Research Institute, University of Michigan , Ann Arbor , USA.,b Department of Otolaryngology , Hearing Rehabilitation Center, University of Michigan , Ann Arbor , USA
| | - Teresa A Zwolan
- b Department of Otolaryngology , Hearing Rehabilitation Center, University of Michigan , Ann Arbor , USA
| | - Bryan E Pfingst
- a Department of Otolaryngology , Kresge Hearing Research Institute, University of Michigan , Ann Arbor , USA
| |
Collapse
|
14
|
Gabay Y, Karni A, Banai K. The perceptual learning of time-compressed speech: A comparison of training protocols with different levels of difficulty. PLoS One 2017; 12:e0176488. [PMID: 28545039 PMCID: PMC5436740 DOI: 10.1371/journal.pone.0176488] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2016] [Accepted: 04/11/2017] [Indexed: 11/18/2022] Open
Abstract
Speech perception can improve substantially with practice (perceptual learning) even in adults. Here we compared the effects of four training protocols that differed in whether and how task difficulty was changed during a training session, in terms of the gains attained and the ability to apply (transfer) these gains to previously un-encountered items (tokens) and to different talkers. Participants trained in judging the semantic plausibility of sentences presented as time-compressed speech and were tested on their ability to reproduce, in writing, the target sentences; trail-by-trial feedback was afforded in all training conditions. In two conditions task difficulty (low or high compression) was kept constant throughout the training session, whereas in the other two conditions task difficulty was changed in an adaptive manner (incrementally from easy to difficult, or using a staircase procedure). Compared to a control group (no training), all four protocols resulted in significant post-training improvement in the ability to reproduce the trained sentences accurately. However, training in the constant-high-compression protocol elicited the smallest gains in deciphering and reproducing trained items and in reproducing novel, untrained, items after training. Overall, these results suggest that training procedures that start off with relatively little signal distortion (“easy” items, not far removed from standard speech) may be advantageous compared to conditions wherein severe distortions are presented to participants from the very beginning of the training session.
Collapse
Affiliation(s)
- Yafit Gabay
- Department of Communications Sciences and Disorders, University of Haifa, Haifa, Israel
- Edmond J. Safra Brain Research Center for the Study of Learning Disabilities, Department of Learning Disabilities, University of Haifa, Haifa, Israel
- Department of Special Education, University of Haifa, Haifa, Israel
- * E-mail:
| | - Avi Karni
- Edmond J. Safra Brain Research Center for the Study of Learning Disabilities, Department of Learning Disabilities, University of Haifa, Haifa, Israel
- Sagol Department of Neurobiology, University of Haifa, Haifa, Israel
| | - Karen Banai
- Department of Communications Sciences and Disorders, University of Haifa, Haifa, Israel
| |
Collapse
|
15
|
Meister H, Fürsen K, Streicher B, Lang-Roth R, Walger M. The Use of Voice Cues for Speaker Gender Recognition in Cochlear Implant Recipients. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2016; 59:546-556. [PMID: 27135985 DOI: 10.1044/2015_jslhr-h-15-0128] [Citation(s) in RCA: 29] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/08/2015] [Accepted: 09/23/2015] [Indexed: 06/05/2023]
Abstract
PURPOSE The focus of this study was to examine the influence of fundamental frequency (F0) and vocal tract length (VTL) modifications on speaker gender recognition in cochlear implant (CI) recipients for different stimulus types. METHOD Single words and sentences were manipulated using isolated or combined F0 and VTL cues. Using an 11-point rating scale, CI recipients and listeners with normal hearing rated the maleness/femaleness of the corresponding voice. RESULTS Speaker gender ratings for combined F0 and VTL modifications were similar across all stimulus types in both CI recipients and listeners with normal hearing, although the CI recipients showed a somewhat larger ambiguity. In contrast to listeners with normal hearing, F0-VTL and F0-only modifications revealed similar ratings in the CI recipients when using words as stimuli. However, when sentences were used, a difference was found between F0-VTL-based and F0-based ratings. Modifying VTL cues alone did not affect ratings in the CI group. CONCLUSIONS Whereas speaker gender ratings by listeners with normal hearing relied on combined VTL and F0 cues, CI recipients made only limited use of VTL cues, which might be one reason behind problems with identifying the speaker on the basis of voice. However, use of the voice cues depended on stimulus type, with the greater information in sentences allowing a more detailed analysis than single words in both listener groups.
Collapse
|
16
|
Shannon RV. Auditory implant research at the House Ear Institute 1989-2013. Hear Res 2015; 322:57-66. [PMID: 25449009 PMCID: PMC4380593 DOI: 10.1016/j.heares.2014.11.003] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/11/2014] [Revised: 11/04/2014] [Accepted: 11/07/2014] [Indexed: 11/29/2022]
Abstract
The House Ear Institute (HEI) had a long and distinguished history of auditory implant innovation and development. Early clinical innovations include being one of the first cochlear implant (CI) centers, being the first center to implant a child with a cochlear implant in the US, developing the auditory brainstem implant, and developing multiple surgical approaches and tools for Otology. This paper reviews the second stage of auditory implant research at House - in-depth basic research on perceptual capabilities and signal processing for both cochlear implants and auditory brainstem implants. Psychophysical studies characterized the loudness and temporal perceptual properties of electrical stimulation as a function of electrical parameters. Speech studies with the noise-band vocoder showed that only four bands of tonotopically arrayed information were sufficient for speech recognition, and that most implant users were receiving the equivalent of 8-10 bands of information. The noise-band vocoder allowed us to evaluate the effects of the manipulation of the number of bands, the alignment of the bands with the original tonotopic map, and distortions in the tonotopic mapping, including holes in the neural representation. Stimulation pulse rate was shown to have only a small effect on speech recognition. Electric fields were manipulated in position and sharpness, showing the potential benefit of improved tonotopic selectivity. Auditory training shows great promise for improving speech recognition for all patients. And the Auditory Brainstem Implant was developed and improved and its application expanded to new populations. Overall, the last 25 years of research at HEI helped increase the basic scientific understanding of electrical stimulation of hearing and contributed to the improved outcomes for patients with the CI and ABI devices. This article is part of a Special Issue entitled .
Collapse
Affiliation(s)
- Robert V Shannon
- Department of Otolaryngology, University of Southern California, Keck School of Medicine of USC, 806 W. Adams Blvd, Los Angeles, CA 90007-2505, USA.
| |
Collapse
|
17
|
Grasmeder ML, Verschuur CA, Batty VB. Optimizing frequency-to-electrode allocation for individual cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:3313. [PMID: 25480076 DOI: 10.1121/1.4900831] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/20/2023]
Abstract
Individual adjustment of frequency-to-electrode assignment in cochlear implants (CIs) may potentially improve speech perception outcomes. Twelve adult CI users were recruited for an experiment, in which frequency maps were adjusted using insertion angles estimated from post-operative x rays; results were analyzed for ten participants with good quality x rays. The allocations were a mapping to the Greenwood function, a compressed map limited to the area containing spiral ganglion (SG) cells, a reduced frequency range map (RFR), and participants' clinical maps. A trial period of at least six weeks was given for the clinical, Greenwood, and SG maps although participants could return to their clinical map if they wished. Performance with the Greenwood map was poor for both sentence and vowel perception and correlated with insertion angle; performance with the SG map was poorer than for the clinical map. The RFR map was significantly better than the clinical map for three participants, for sentence perception, but worse for three others. Those with improved performance had relatively deep insertions and poor electrode discrimination ability for apical electrodes. The results suggest that CI performance could be improved by adjustment of the frequency allocation, based on a measure of insertion angle and/or electrode discrimination ability.
Collapse
Affiliation(s)
- Mary L Grasmeder
- Auditory Implant Service, Faculty of Engineering and the Environment, Building 19, University of Southampton, Southampton SO17 1BJ, United Kingdom
| | - Carl A Verschuur
- Auditory Implant Service, Faculty of Engineering and the Environment, Building 19, University of Southampton, Southampton SO17 1BJ, United Kingdom
| | - Vincent B Batty
- Wessex Neurological Centre, University Hospital Southampton NHS Foundation Trust, Tremona Road, Southampton SO16 6YD, United Kingdom
| |
Collapse
|
18
|
Svirsky MA, Talavage TM, Sinha S, Neuburger H, Azadpour M. Gradual adaptation to auditory frequency mismatch. Hear Res 2014; 322:163-70. [PMID: 25445816 DOI: 10.1016/j.heares.2014.10.008] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/13/2014] [Revised: 10/13/2014] [Accepted: 10/16/2014] [Indexed: 12/01/2022]
Abstract
What is the best way to help humans adapt to a distorted sensory input? Interest in this question is more than academic. The answer may help facilitate auditory learning by people who became deaf after learning language and later received a cochlear implant (a neural prosthesis that restores hearing through direct electrical stimulation of the auditory nerve). There is evidence that some cochlear implants (which provide information that is spectrally degraded to begin with) stimulate neurons with higher characteristic frequency than the acoustic frequency of the original stimulus. In other words, the stimulus is shifted in frequency with respect to what the listener expects to hear. This frequency misalignment may have a negative influence on speech perception by CI users. However, a perfect frequency-place alignment may result in the loss of important low frequency speech information. A trade-off may involve a gradual approach: start with correct frequency-place alignment to allow listeners to adapt to the spectrally degraded signal first, and then gradually increase the frequency shift to allow them to adapt to it over time. We used an acoustic model of a cochlear implant to measure adaptation to a frequency-shifted signal, using either the gradual approach or the "standard" approach (sudden imposition of the frequency shift). Listeners in both groups showed substantial auditory learning, as measured by increases in speech perception scores over the course of fifteen one-hour training sessions. However, the learning process was faster for listeners who were exposed to the gradual approach. These results suggest that gradual rather than sudden exposure may facilitate perceptual learning in the face of a spectrally degraded, frequency-shifted input. This article is part of a Special Issue entitled <Lasker Award>.
Collapse
Affiliation(s)
- Mario A Svirsky
- Dept. of Otolaryngology-HNS, New York University School of Medicine, New York, NY, USA; Center of Neural Science, New York University, New York, NY, USA.
| | - Thomas M Talavage
- ECE, Purdue University, West Lafayette, IN, USA; BME Depts., Purdue University, West Lafayette, IN, USA
| | | | - Heidi Neuburger
- Dept. of Otolaryngology-HNS, Indiana University School of Medicine, Indianapolis, IN, USA
| | - Mahan Azadpour
- Dept. of Otolaryngology-HNS, New York University School of Medicine, New York, NY, USA
| |
Collapse
|
19
|
Dickinson AM, Baker R, Siciliano C, Munro KJ. Adaptation to nonlinear frequency compression in normal-hearing adults: a comparison of training approaches. Int J Audiol 2014; 53:719-29. [PMID: 24975233 DOI: 10.3109/14992027.2014.921338] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
OBJECTIVE To identify which training approach, if any, is most effective for improving perception of frequency-compressed speech. DESIGN A between-subject design using repeated measures. STUDY SAMPLE Forty young adults with normal hearing were randomly allocated to one of four groups: a training group (sentence or consonant) or a control group (passive exposure or test-only). Test and training material differed in terms of material and speaker. RESULTS On average, sentence training and passive exposure led to significantly improved sentence recognition (11.0% and 11.7%, respectively) compared with the consonant training group (2.5%) and test-only group (0.4%), whilst, consonant training led to significantly improved consonant recognition (8.8%) compared with the sentence training group (1.9%), passive exposure group (2.8%), and test-only group (0.8%). CONCLUSIONS Sentence training led to improved sentence recognition, whilst consonant training led to improved consonant recognition. This suggests learning transferred between speakers and material but not stimuli. Passive exposure to sentence material led to an improvement in sentence recognition that was equivalent to gains from active training. This suggests that it may be possible to adapt passively to frequency-compressed speech.
Collapse
|
20
|
Guediche S, Blumstein SE, Fiez JA, Holt LL. Speech perception under adverse conditions: insights from behavioral, computational, and neuroscience research. Front Syst Neurosci 2014; 7:126. [PMID: 24427119 PMCID: PMC3879477 DOI: 10.3389/fnsys.2013.00126] [Citation(s) in RCA: 47] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2013] [Accepted: 12/16/2013] [Indexed: 01/06/2023] Open
Abstract
Adult speech perception reflects the long-term regularities of the native language, but it is also flexible such that it accommodates and adapts to adverse listening conditions and short-term deviations from native-language norms. The purpose of this article is to examine how the broader neuroscience literature can inform and advance research efforts in understanding the neural basis of flexibility and adaptive plasticity in speech perception. Specifically, we highlight the potential role of learning algorithms that rely on prediction error signals and discuss specific neural structures that are likely to contribute to such learning. To this end, we review behavioral studies, computational accounts, and neuroimaging findings related to adaptive plasticity in speech perception. Already, a few studies have alluded to a potential role of these mechanisms in adaptive plasticity in speech perception. Furthermore, we consider research topics in neuroscience that offer insight into how perception can be adaptively tuned to short-term deviations while balancing the need to maintain stability in the perception of learned long-term regularities. Consideration of the application and limitations of these algorithms in characterizing flexible speech perception under adverse conditions promises to inform theoretical models of speech.
Collapse
Affiliation(s)
- Sara Guediche
- Department of Cognitive, Linguistic, and Psychological Sciences, Brown UniversityProvidence, RI, USA
| | - Sheila E. Blumstein
- Department of Cognitive, Linguistic, and Psychological Sciences, Brown UniversityProvidence, RI, USA
- Department of Cognitive, Linguistic, and Psychological Sciences, Brain Institute, Brown UniversityProvidence, RI, USA
| | - Julie A. Fiez
- Department of Neuroscience, Center for Neuroscience at the University of Pittsburgh, University of PittsburghPittsburgh, PA, USA
- Department of Psychology, University of PittsburghPittsburgh, PA, USA
- Department of Psychology at Carnegie Mellon University and Department of Neuroscience at the University of Pittsburgh, Center for the Neural Basis of CognitionPittsburgh, PA, USA
| | - Lori L. Holt
- Department of Neuroscience, Center for Neuroscience at the University of Pittsburgh, University of PittsburghPittsburgh, PA, USA
- Department of Psychology at Carnegie Mellon University and Department of Neuroscience at the University of Pittsburgh, Center for the Neural Basis of CognitionPittsburgh, PA, USA
- Department of Psychology, Carnegie Mellon UniversityPittsburgh, PA, USA
| |
Collapse
|
21
|
Li T, Fu QJ. Perceptual adaptation of voice gender discrimination with spectrally shifted vowels. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2011; 54:1240-1245. [PMID: 21173392 PMCID: PMC3580211 DOI: 10.1044/1092-4388(2010/10-0168)] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
PURPOSE To determine whether perceptual adaptation improves voice gender discrimination of spectrally shifted vowels and, if so, which acoustic cues contribute to the improvement. METHOD Voice gender discrimination was measured for 10 normal-hearing subjects, during 5 days of adaptation to spectrally shifted vowels, produced by processing the speech of 5 male and 5 female talkers with 16-channel sine-wave vocoders. The subjects were randomly divided into 2 groups; one subjected to 50-Hz, and the other to 200-Hz, temporal envelope cutoff frequencies. No preview or feedback was provided. RESULTS There was significant adaptation in voice gender discrimination with the 200-Hz cutoff frequency, but significant improvement was observed only for 3 female talkers with F(0) > 180 Hz and 3 male talkers with F(0) < 170 Hz. There was no significant adaptation with the 50-Hz cutoff frequency. CONCLUSIONS Temporal envelope cues are important for voice gender discrimination under spectral shift conditions with perceptual adaptation, but spectral shift may limit the exclusive use of spectral information and/or the use of formant structure on voice gender discrimination. The results have implications for cochlear implant users and for understanding voice gender discrimination.
Collapse
Affiliation(s)
- Tianhao Li
- House Ear Institute, Los Angeles, CA, USA.
| | | |
Collapse
|
22
|
Li T, Fu QJ. Voice gender discrimination provides a measure of more than pitch-related perception in cochlear implant users. Int J Audiol 2011; 50:498-502. [PMID: 21696330 DOI: 10.3109/14992027.2011.576274] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022]
Abstract
OBJECTIVES (1) To investigate whether voice gender discrimination (VGD) could be a useful indicator of the spectral and temporal processing abilities of individual cochlear implant (CI) users; (2) To examine the relationship between VGD and speech recognition with CI when comparable acoustic cues are used for both perception processes. DESIGN VGD was measured using two talker sets with different inter-gender fundamental frequencies (F(0)), as well as different acoustic CI simulations. Vowel and consonant recognition in quiet and noise were also measured and compared with VGD performance. STUDY SAMPLE Eleven postlingually deaf CI users. RESULTS The results showed that (1) mean VGD performance differed for different stimulus sets, (2) VGD and speech recognition performance varied among individual CI users, and (3) individual VGD performance was significantly correlated with speech recognition performance under certain conditions. CONCLUSIONS VGD measured with selected stimulus sets might be useful for assessing not only pitch-related perception, but also spectral and temporal processing by individual CI users. In addition to improvements in spectral resolution and modulation detection, the improvement in higher modulation frequency discrimination might be particularly important for CI users in noisy environments.
Collapse
Affiliation(s)
- Tianhao Li
- Division of Communication and Auditory Neuroscience, House Ear Institute, 2100 West Third Street, Los Angeles, CA 90057, USA.
| | | |
Collapse
|
23
|
Li T, Fu QJ. Effects of spectral shifting on speech perception in noise. Hear Res 2010; 270:81-8. [PMID: 20868733 DOI: 10.1016/j.heares.2010.09.005] [Citation(s) in RCA: 32] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/03/2009] [Revised: 09/08/2010] [Accepted: 09/14/2010] [Indexed: 11/16/2022]
Abstract
The present study used eight normal-hearing (NH) subjects, listening to acoustic cochlear implant (CI) simulations, to examine the effects of spectral shifting on speech recognition in noise. Speech recognition was measured using spectrally matched and shifted speech (vowels, consonants, and IEEE sentences), generated by 8-channel, sine-wave vocoder. Measurements were made in quiet and in noise (speech-shaped static noise and speech-babble at 5 dB signal-to-noise ratio). One spectral match condition and four spectral shift conditions were investigated: 2 mm, 3 mm, and 4 mm linear shift, and 3 mm shift with compression, in terms of cochlear distance. Results showed that speech recognition scores dropped because of noise and spectral shifting, and that the interactive effects of spectral shifting and background conditions depended on the degree/type of spectral shift, background conditions, and the speech test materials. There was no significant interaction between spectral shifting and two noise conditions for all speech test materials. However, significant interactions between linear spectral shifts and all background conditions were found in sentence recognition; significant interactions between spectral shift types and all background conditions were found in vowel recognition. Overall, the results suggest that tonotopic mismatch may affect performance of CI users in complex listening environments.
Collapse
Affiliation(s)
- Tianhao Li
- Division of Communication and Auditory Neuroscience, House Ear Institute, Los Angeles, CA 90057, USA.
| | | |
Collapse
|
24
|
Siciliano CM, Faulkner A, Rosen S, Mair K. Resistance to learning binaurally mismatched frequency-to-place maps: implications for bilateral stimulation with cochlear implants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 127:1645-60. [PMID: 20329863 DOI: 10.1121/1.3293002] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Simulations of monaural cochlear implants in normal hearing listeners have shown that the deleterious effects of upward spectral shifting on speech perception can be overcome with training. This study simulates bilateral stimulation with a unilateral spectral shift to investigate whether listeners can adapt to upward-shifted speech information presented together with contralateral unshifted information. A six-channel, dichotic, interleaved sine-carrier vocoder simulated a binaurally mismatched frequency-to-place map. Odd channels were presented to one ear with an upward frequency shift equivalent to 6 mm on the basilar membrane, while even channels were presented to the contralateral ear unshifted. In Experiment 1, listeners were trained for 5.3 h with either the binaurally mismatched processor or with just the shifted monaural bands. In Experiment 2, the duration of training was 10 h, and the trained condition alternated between those of Experiment 1. While listeners showed learning in both experiments, intelligibility with the binaurally mismatched processor never exceeded, intelligibility with just the three unshifted bands, suggesting that listeners did not benefit from combining the mismatched maps, even though there was clear scope to do so. Frequency-place map alignment may thus be of importance when optimizing bilateral devices of the type studied here.
Collapse
Affiliation(s)
- Catherine M Siciliano
- Speech, Hearing and Phonetic Sciences, Division of Psychology and Language Sciences, UCL, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom.
| | | | | | | |
Collapse
|