1
|
Guest DR, Rajappa N, Oxenham AJ. Limitations in human auditory spectral analysis at high frequencies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 156:326-340. [PMID: 38990035 PMCID: PMC11240212 DOI: 10.1121/10.0026475] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 06/07/2024] [Indexed: 07/12/2024]
Abstract
Humans are adept at identifying spectral patterns, such as vowels, in different rooms, at different sound levels, or produced by different talkers. How this feat is achieved remains poorly understood. Two psychoacoustic analogs of spectral pattern recognition are spectral profile analysis and spectrotemporal ripple direction discrimination. This study tested whether pattern-recognition abilities observed previously at low frequencies are also observed at extended high frequencies. At low frequencies (center frequency ∼500 Hz), listeners were able to achieve accurate profile-analysis thresholds, consistent with prior literature. However, at extended high frequencies (center frequency ∼10 kHz), listeners' profile-analysis thresholds were either unmeasurable or could not be distinguished from performance based on overall loudness cues. A similar pattern of results was observed with spectral ripple discrimination, where performance was again considerably better at low than at high frequencies. Collectively, these results suggest a severe deficit in listeners' ability to analyze patterns of intensity across frequency in the extended high-frequency region that cannot be accounted for by cochlear frequency selectivity. One interpretation is that the auditory system is not optimized to analyze such fine-grained across-frequency profiles at extended high frequencies, as they are not typically informative for everyday sounds.
Collapse
Affiliation(s)
- Daniel R Guest
- Department of Biomedical Engineering, University of Rochester, Rochester, New York 14642, USA
| | - Neha Rajappa
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Andrew J Oxenham
- Department of Psychology, University of Minnesota, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
2
|
Cychosz M, Winn MB, Goupell MJ. How to vocode: Using channel vocoders for cochlear-implant research. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2407-2437. [PMID: 38568143 PMCID: PMC10994674 DOI: 10.1121/10.0025274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Revised: 02/14/2024] [Accepted: 02/23/2024] [Indexed: 04/05/2024]
Abstract
The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, California 90095, USA
| | - Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, College Park, Maryland 20742, USA
| |
Collapse
|
3
|
Creff G, Lambert C, Coudert P, Pean V, Laurent S, Godey B. Comparison of Tonotopic and Default Frequency Fitting for Speech Understanding in Noise in New Cochlear Implantees: A Prospective, Randomized, Double-Blind, Cross-Over Study. Ear Hear 2024; 45:35-52. [PMID: 37823850 DOI: 10.1097/aud.0000000000001423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/13/2023]
Abstract
OBJECTIVES While cochlear implants (CIs) have provided benefits for speech recognition in quiet for subjects with severe-to-profound hearing loss, speech recognition in noise remains challenging. A body of evidence suggests that reducing frequency-to-place mismatch may positively affect speech perception. Thus, a fitting method based on a tonotopic map may improve speech perception results in quiet and noise. The aim of our study was to assess the impact of a tonotopic map on speech perception in noise and quiet in new CI users. DESIGN A prospective, randomized, double-blind, two-period cross-over study in 26 new CI users was performed over a 6-month period. New CI users older than 18 years with bilateral severe-to-profound sensorineural hearing loss or complete hearing loss for less than 5 years were selected in the University Hospital Centre of Rennes in France. An anatomical tonotopic map was created using postoperative flat-panel computed tomography and a reconstruction software based on the Greenwood function. Each participant was randomized to receive a conventional map followed by a tonotopic map or vice versa. Each setting was maintained for 6 weeks, at the end of which participants performed speech perception tasks. The primary outcome measure was speech recognition in noise. Participants were allocated to sequences by block randomization of size two with a ratio 1:1 (CONSORT Guidelines). Participants and those assessing the outcomes were blinded to the intervention. RESULTS Thirteen participants were randomized to each sequence. Two of the 26 participants recruited (one in each sequence) had to be excluded due to the COVID-19 pandemic. Twenty-four participants were analyzed. Speech recognition in noise was significantly better with the tonotopic fitting at all signal-to-noise ratio (SNR) levels tested [SNR = +9 dB, p = 0.002, mean effect (ME) = 12.1%, 95% confidence interval (95% CI) = 4.9 to 19.2, standardized effect size (SES) = 0.71; SNR = +6 dB, p < 0.001, ME = 16.3%, 95% CI = 9.8 to 22.7, SES = 1.07; SNR = +3 dB, p < 0.001 ME = 13.8%, 95% CI = 6.9 to 20.6, SES = 0.84; SNR = 0 dB, p = 0.003, ME = 10.8%, 95% CI = 4.1 to 17.6, SES = 0.68]. Neither period nor interaction effects were observed for any signal level. Speech recognition in quiet ( p = 0.66) and tonal audiometry ( p = 0.203) did not significantly differ between the two settings. 92% of the participants kept the tonotopy-based map after the study period. No correlation was found between speech-in-noise perception and age, duration of hearing deprivation, angular insertion depth, or position or width of the frequency filters allocated to the electrodes. CONCLUSION For new CI users, tonotopic fitting appears to be more efficient than the default frequency fitting because it allows for better speech recognition in noise without compromising understanding in quiet.
Collapse
Affiliation(s)
- Gwenaelle Creff
- Department of Otolaryngology-Head and Neck Surgery (HNS), University Hospital, Rennes, France
- MediCIS, LTSI (Image and Signal Processing Laboratory), INSERM, U1099, Rennes, France
| | - Cassandre Lambert
- Department of Otolaryngology-Head and Neck Surgery (HNS), University Hospital, Rennes, France
| | - Paul Coudert
- Department of Otolaryngology-Head and Neck Surgery (HNS), University Hospital, Rennes, France
| | | | | | - Benoit Godey
- Department of Otolaryngology-Head and Neck Surgery (HNS), University Hospital, Rennes, France
- MediCIS, LTSI (Image and Signal Processing Laboratory), INSERM, U1099, Rennes, France
- Hearing Aid Academy, Javene, France
| |
Collapse
|
4
|
Dillon MT, Helpard L, Brown KD, Selleck AM, Richter ME, Rooth MA, Thompson NJ, Dedmon MM, Ladak HM, Agrawal S. Influence of the Frequency-to-Place Function on Recognition with Place-Based Cochlear Implant Maps. Laryngoscope 2023; 133:3540-3547. [PMID: 37078508 DOI: 10.1002/lary.30710] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Revised: 04/07/2023] [Accepted: 04/11/2023] [Indexed: 04/21/2023]
Abstract
OBJECTIVE Comparison of acute speech recognition for cochlear implant (CI) alone and electric-acoustic stimulation (EAS) users listening with default maps or place-based maps using either a spiral ganglion (SG) or a new Synchrotron Radiation-Artificial Intelligence (SR-AI) frequency-to-place function. METHODS Thirteen adult CI-alone or EAS users completed a task of speech recognition at initial device activation with maps that differed in the electric filter frequency assignments. The three map conditions were: (1) maps with the default filter settings (default map), (2) place-based maps with filters aligned to cochlear SG tonotopicity using the SG function (SG place-based map), and (3) place-based maps with filters aligned to cochlear Organ of Corti (OC) tonotopicity using the SR-AI function (SR-AI place-based map). Speech recognition was evaluated using a vowel recognition task. Performance was scored as the percent correct for formant 1 recognition due to the rationale that the maps would deviate the most in the estimated cochlear place frequency for low frequencies. RESULTS On average, participants had better performance with the OC SR-AI place-based map as compared to the SG place-based map and the default map. A larger performance benefit was observed for EAS users than for CI-alone users. CONCLUSION These pilot data suggest that EAS and CI-alone users may experience better performance with a patient-centered mapping approach that accounts for the variability in cochlear morphology (OC SR-AI frequency-to-place function) in the individualization of the electric filter frequencies (place-based mapping procedure). LEVEL OF EVIDENCE 3 Laryngoscope, 133:3540-3547, 2023.
Collapse
Affiliation(s)
- Margaret T Dillon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Luke Helpard
- School of Biomedical Engineering, Western University, London, Ontario, Canada
| | - Kevin D Brown
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - A Morgan Selleck
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Margaret E Richter
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Meredith A Rooth
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Nicholas J Thompson
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Matthew M Dedmon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, USA
| | - Hanif M Ladak
- School of Biomedical Engineering, Western University, London, Ontario, Canada
- Department of Electrical and Computer Engineering, Western University, London, Ontario, Canada
- Department of Medical Biophysics, Western University, London, Ontario, Canada
- Department of Otolaryngology - Head & Neck Surgery, Western University, London, Ontario, Canada
| | - Sumit Agrawal
- Department of Otolaryngology - Head & Neck Surgery, Western University, London, Ontario, Canada
| |
Collapse
|
5
|
Dillon MT, Buss E, Johnson AD, Canfarotta MW, O’Connell BP. Comparison of Two Place-Based Mapping Procedures on Masked Sentence Recognition as a Function of Electrode Array Angular Insertion Depth and Presence of Acoustic Low-Frequency Information: A Simulation Study. Audiol Neurootol 2023; 28:478-487. [PMID: 37482054 PMCID: PMC10948008 DOI: 10.1159/000531262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2022] [Accepted: 05/23/2023] [Indexed: 07/25/2023] Open
Abstract
INTRODUCTION Cochlear implant (CI) and electric-acoustic stimulation (EAS) users may experience better performance with maps that align the electric filter frequencies to the cochlear place frequencies, known as place-based maps, than with maps that present spectrally shifted information. Individual place-based mapping procedures differ in the frequency content that is aligned to cochlear tonotopicity versus discarded or spectrally shifted. The performance benefit with different place-based maps may vary due to individual differences in angular insertion depth (AID) of the electrode array and whether functional acoustic low-frequency information is available in the implanted ear. The present study compared masked speech recognition with two types of place-based maps as a function of AID and presence of acoustic low-frequency information. METHODS Sixty adults with normal hearing listened acutely to CI or EAS simulations of two types of place-based maps for one of three cases of electrode arrays at shallow AIDs. The strict place-based (Strict-PB) map aligned the low- and mid-frequency information to cochlear tonotopicity and discarded information below the frequency associated with the most apical electrode contact. The alternative place-based map (LFshift-PB) aligned the mid-frequency information to cochlear tonotopicity and provided more of the speech spectrum by compressing low-frequency information on the apical electrode contacts (i.e., <1 kHz). Three actual cases of a 12-channel, 24-mm electrode array were simulated by assigning the carrier frequency for an individual channel as the cochlear place frequency of the associated electrode contact. The AID and cochlear place frequency for the most apical electrode contact were 460° and 498 Hz for case 1, 389° and 728 Hz for case 2, and 335° and 987 Hz for case 3, respectively. RESULTS Generally, better performance was observed with the Strict-PB maps for cases 1 and 2, where mismatches were 2-4 octaves for the most apical channel with the LFshift-PB map. Similar performance was observed between maps for case 3. For the CI simulations, performance with the Strict-PB map declined with decreases in AID, while performance with the LFshift-PB map remained stable across cases. For the EAS simulations, performance with the Strict-PB map remained stable across cases, while performance with the LFshift-PB map improved with decreases in AID. CONCLUSIONS Listeners demonstrated differences with the Strict-PB versus LFshift-PB maps as a function of AID and whether acoustic low-frequency information was available (CI vs. EAS). These data support the use of the Strict-PB mapping procedure for AIDs ≥335°, though further study including time for acclimatization in CI and EAS users is warranted.
Collapse
Affiliation(s)
- Margaret T. Dillon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Emily Buss
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Alec D. Johnson
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Michael W. Canfarotta
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Brendan P. O’Connell
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
- Charlotte Eye Ear Nose & Throat Associates, P.A., Charlotte, NC, USA
| |
Collapse
|
6
|
Cleary M, DeRoy Milvae K, Nguyen N, Bernstein JGW, Goupell MJ. Effect of experimentally introduced interaural frequency mismatch on sentence recognition in bilateral cochlear-implant listeners. JASA EXPRESS LETTERS 2023; 3:044401. [PMID: 37096891 PMCID: PMC10080388 DOI: 10.1121/10.0017705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Accepted: 03/13/2023] [Indexed: 05/03/2023]
Abstract
Bilateral cochlear-implant users experience interaural frequency mismatch because of asymmetries in array insertion and frequency-to-electrode assignment. To explore the acute perceptual consequences of such mismatch, sentence recognition in quiet was measured in nine bilateral cochlear-implant listeners as frequency allocations in the poorer ear were shifted by ±1.5, ±3, and ±4.5 mm using experimental programs. Shifts in frequency allocation >3 mm reduced bilateral sentence scores below those for the better ear alone, suggesting that the poorer ear interfered with better-ear perception. This was not a result of fewer active channels; deactivating electrodes without frequency shifting had minimal effect.
Collapse
Affiliation(s)
- Miranda Cleary
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Kristina DeRoy Milvae
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Nicole Nguyen
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Joshua G W Bernstein
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, Maryland 20889, , , , ,
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
7
|
Cleary M, DeRoy Milvae K, Nguyen N, Bernstein JGW, Goupell MJ. Effect of experimentally introduced interaural frequency mismatch on sentence recognition in bilateral cochlear-implant listeners. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2023:2023.01.06.23284274. [PMID: 36711489 PMCID: PMC9882401 DOI: 10.1101/2023.01.06.23284274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/11/2023]
Abstract
Bilateral cochlear-implant users experience interaural frequency mismatch because of asymmetries in array insertion and frequency-to-electrode assignment. To explore the acute perceptual consequences of such mismatch, sentence recognition in quiet was measured in nine bilateral cochlear-implant listeners as frequency allocations in the poorer ear were shifted by ±1.5, ±3 and ±4.5 mm using experimental programs. Shifts in frequency allocation >3 mm were found to reduce bilateral sentence scores below those for the better ear alone, suggesting that the poorer ear interfered with better-ear perception. This was not a result of fewer active channels; deactivating electrodes without frequency shifting had minimal effect.
Collapse
Affiliation(s)
- Miranda Cleary
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Kristina DeRoy Milvae
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA,Department of Communicative Disorders and Sciences, University at Buffalo, Buffalo, NY, USA
| | - Nicole Nguyen
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Joshua G. W. Bernstein
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD, USA
| | - Matthew J. Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| |
Collapse
|
8
|
Association between lateral wall electrode array insertion parameters and audiological outcomes in bilateral cochlear implantation. Eur Arch Otorhinolaryngol 2022; 280:2707-2714. [PMID: 36436080 PMCID: PMC10175364 DOI: 10.1007/s00405-022-07756-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2022] [Accepted: 11/15/2022] [Indexed: 11/28/2022]
Abstract
Abstract
Purpose
The aims of this study were to compare speech recognition at different postoperative times for both ears in bilaterally implanted patients and to assess the influence of the time of deafness, frequency-to-place mismatch, angular insertion depth (AID) and angular separation between neighbouring electrode contacts on audiometric outcomes.
Methods
This study was performed at an academic tertiary referral centre. A total of 19 adult patients (6 men, 13 women), who received sequential bilateral implantation with lateral wall electrode arrays, were analysed in retrospective. Statistical analysis was performed using two-sided t test, Wilcoxon test, median test, and Spearman’s correlation.
Results
Postlingually deafened patients (deafness after the age of 10) had a significantly better speech perception (WRS65[CI]) than the perilingually deafened subjects (deafness at the age of 1–10 years) (p < 0.001). Comparison of cochlear duct length between peri- and postlingually deafened subjects showed a slightly significantly smaller cochleae in perilingual patients (p = 0.045). No association between frequency-to-place mismatch as well as angular separation and speech perception could be detected. There was even no significant difference between the both ears in the intraindividual comparison, even if insertion parameters differed.
Conclusion
The exact electrode position seems to have less influence on the speech comprehension of CI patients than already established parameters as preoperative speech recognition or duration of deafness.
Collapse
|
9
|
American Cochlear Implant Alliance Task Force Guidelines for Clinical Assessment and Management of Adult Cochlear Implantation for Single-Sided Deafness. Ear Hear 2022; 43:1605-1619. [PMID: 35994570 PMCID: PMC9592177 DOI: 10.1097/aud.0000000000001260] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
The indications for cochlear implantation have expanded to include individuals with profound sensorineural hearing loss in the impaired ear and normal hearing (NH) in the contralateral ear, known as single-sided deafness (SSD). There are additional considerations for the clinical assessment and management of adult cochlear implant candidates and recipients with SSD as compared to conventional cochlear implant candidates with bilateral moderate to profound sensorineural hearing loss. The present report reviews the current evidence relevant to the assessment and management of adults with SSD. A systematic review was also conducted on published studies that investigated outcomes of cochlear implant use on measures of speech recognition in quiet and noise, sound source localization, tinnitus perception, and quality of life for this patient population. Expert consensus and systematic review of the current literature were combined to provide guidance for the clinical assessment and management of adults with SSD.
Collapse
|
10
|
Ananthakrishnan S, Luo X. Effects of Temporal Envelope Cutoff Frequency, Number of Channels, and Carrier Type on Brainstem Neural Representation of Pitch in Vocoded Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:3146-3164. [PMID: 35944032 DOI: 10.1044/2022_jslhr-21-00576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
PURPOSE The objective of this study was to determine if and how the subcortical neural representation of pitch cues in listeners with normal hearing is affected by systematic manipulation of vocoder parameters. METHOD This study assessed the effects of temporal envelope cutoff frequency (50 and 500 Hz), number of channels (1-32), and carrier type (sine-wave and noise-band) on brainstem neural representation of fundamental frequency (f o) in frequency-following responses (FFRs) to vocoded vowels of 15 young adult listeners with normal hearing. RESULTS Results showed that FFR f o strength (quantified as absolute f o magnitude divided by noise floor [NF] magnitude) significantly improved with 500-Hz vs. 50-Hz temporal envelopes for all channel numbers and both carriers except the 1-channel noise-band vocoder. FFR f o strength with 500-Hz temporal envelopes significantly improved when the channel number increased from 1 to 2, but it either declined (sine-wave vocoders) or saturated (noise-band vocoders) when the channel number increased from 4 to 32. FFR f o strength with 50-Hz temporal envelopes was similarly small for both carriers with all channel numbers, except for a significant improvement with the 16-channel sine-wave vocoder. With 500-Hz temporal envelopes, FFR f o strength was significantly greater for sine-wave vocoders than for noise-band vocoders with channel numbers 1-8; no significant differences were seen with 16 and 32 channels. With 50-Hz temporal envelopes, the carrier effect was only observed with 16 channels. In contrast, there was no significant carrier effect for the absolute f o magnitude. Compared to sine-wave vocoders, noise-band vocoders had a higher NF and thus lower relative FFR f o strength. CONCLUSIONS It is important to normalize the f o magnitude relative to the NF when analyzing the FFRs to vocoded speech. The physiological findings reported here may result from the availability of f o-related temporal periodicity and spectral sidelobes in vocoded signals and should be considered when selecting vocoder parameters and interpreting results in future physiological studies. In general, the dependence of brainstem neural phase-locking strength to f o on vocoder parameters may confound the comparison of pitch-related behavioral results across different vocoder designs.
Collapse
Affiliation(s)
| | - Xin Luo
- Program of Speech and Hearing Science, College of Health Solutions, Arizona State University, Tempe
| |
Collapse
|
11
|
Dillon MT, O'Connell BP, Canfarotta MW, Buss E, Hopfinger J. Effect of Place-Based Versus Default Mapping Procedures on Masked Speech Recognition: Simulations of Cochlear Implant Alone and Electric-Acoustic Stimulation. Am J Audiol 2022; 31:322-337. [PMID: 35394798 DOI: 10.1044/2022_aja-21-00123] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE Cochlear implant (CI) recipients demonstrate variable speech recognition when listening with a CI-alone or electric-acoustic stimulation (EAS) device, which may be due in part to electric frequency-to-place mismatches created by the default mapping procedures. Performance may be improved if the filter frequencies are aligned with the cochlear place frequencies, known as place-based mapping. Performance with default maps versus an experimental place-based map was compared for participants with normal hearing when listening to CI-alone or EAS simulations to observe potential outcomes prior to initiating an investigation with CI recipients. METHOD A noise vocoder simulated CI-alone and EAS devices, mapped with default or place-based procedures. The simulations were based on an actual 24-mm electrode array recipient, whose insertion angles for each electrode contact were used to estimate the respective cochlear place frequency. The default maps used the filter frequencies assigned by the clinical software. The filter frequencies for the place-based maps aligned with the cochlear place frequencies for individual contacts in the low- to mid-frequency cochlear region. For the EAS simulations, low-frequency acoustic information was filtered to simulate aided low-frequency audibility. Performance was evaluated for the AzBio sentences presented in a 10-talker masker at +5 dB signal-to-noise ratio (SNR), +10 dB SNR, and asymptote. RESULTS Performance was better with the place-based maps as compared with the default maps for both CI-alone and EAS simulations. For instance, median performance at +10 dB SNR for the CI-alone simulation was 57% correct for the place-based map and 20% for the default map. For the EAS simulation, those values were 59% and 37% correct. Adding acoustic low-frequency information resulted in a similar benefit for both maps. CONCLUSIONS Reducing frequency-to-place mismatches, such as with the experimental place-based mapping procedure, produces a greater benefit in speech recognition than maximizing bandwidth for CI-alone and EAS simulations. Ongoing work is evaluating the initial and long-term performance benefits in CI-alone and EAS users. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.19529053.
Collapse
Affiliation(s)
- Margaret T. Dillon
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
- Division of Speech and Hearing Sciences, Department of Allied Health Sciences, University of North Carolina at Chapel Hill
| | - Brendan P. O'Connell
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Michael W. Canfarotta
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Emily Buss
- Department of Otolaryngology/Head & Neck Surgery, University of North Carolina at Chapel Hill
| | - Joseph Hopfinger
- Department of Psychology and Neuroscience, University of North Carolina at Chapel Hill
| |
Collapse
|
12
|
Jaekel BN, Weinstein S, Newman RS, Goupell MJ. Impacts of signal processing factors on perceptual restoration in cochlear-implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:2898. [PMID: 35649892 PMCID: PMC9054268 DOI: 10.1121/10.0010258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Cochlear-implant (CI) users have previously demonstrated perceptual restoration, or successful repair of noise-interrupted speech, using the interrupted sentences paradigm [Bhargava, Gaudrain, and Başkent (2014). "Top-down restoration of speech in cochlear-implant users," Hear. Res. 309, 113-123]. The perceptual restoration effect was defined experimentally as higher speech understanding scores with noise-burst interrupted sentences compared to silent-gap interrupted sentences. For the perceptual restoration illusion to occur, it is often necessary for the masking or interrupting noise bursts to have a higher intensity than the adjacent speech signal to be perceived as a plausible masker. Thus, signal processing factors like noise reduction algorithms and automatic gain control could have a negative impact on speech repair in this population. Surprisingly, evidence that participants with cochlear implants experienced the perceptual restoration illusion was not observed across the two planned experiments. A separate experiment, which aimed to provide a close replication of previous work on perceptual restoration in CI users, also found no consistent evidence of perceptual restoration, contrasting the original study's previously reported findings. Typical speech repair of interrupted sentences was not observed in the present work's sample of CI users, and signal-processing factors did not appear to affect speech repair.
Collapse
Affiliation(s)
- Brittany N Jaekel
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Sarah Weinstein
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Rochelle S Newman
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
13
|
Thomas M, Willis S, Galvin JJ, Fu QJ. Effects of tonotopic matching and spatial cues on segregation of competing speech in simulations of bilateral cochlear implants. PLoS One 2022; 17:e0270759. [PMID: 35788202 PMCID: PMC9255761 DOI: 10.1371/journal.pone.0270759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2021] [Accepted: 06/16/2022] [Indexed: 11/18/2022] Open
Abstract
In the clinical fitting of cochlear implants (CIs), the lowest input acoustic frequency is typically much lower than the characteristic frequency associated with the most apical electrode position, due to the limited electrode insertion depth. For bilateral CI users, electrode positions may differ across ears. However, the same acoustic-to-electrode frequency allocation table (FAT) is typically assigned to both ears. As such, bilateral CI users may experience both intra-aural frequency mismatch within each ear and inter-aural mismatch across ears. This inter-aural mismatch may limit the ability of bilateral CI users to take advantage of spatial cues when attempting to segregate competing speech. Adjusting the FAT to tonotopically match the electrode position in each ear (i.e., increasing the low acoustic input frequency) is theorized to reduce this inter-aural mismatch. Unfortunately, this approach may also introduce the loss of acoustic information below the modified input acoustic frequency. The present study explored the trade-off between reduced inter-aural frequency mismatch and low-frequency information loss for segregation of competing speech. Normal-hearing participants were tested while listening to acoustic simulations of bilateral CIs. Speech reception thresholds (SRTs) were measured for target sentences produced by a male talker in the presence of two different male talkers. Masker speech was either co-located with or spatially separated from the target speech. The bilateral CI simulations were produced by 16-channel sinewave vocoders; the simulated insertion depth was fixed in one ear and varied in the other ear, resulting in an inter-aural mismatch of 0, 2, or 6 mm in terms of cochlear place. Two FAT conditions were compared: 1) clinical (200-8000 Hz in both ears), or 2) matched to the simulated insertion depth in each ear. Results showed that SRTs were significantly lower with the matched than with the clinical FAT, regardless of the insertion depth or spatial configuration of the masker speech. The largest improvement in SRTs with the matched FAT was observed when the inter-aural mismatch was largest (6 mm). These results suggest that minimizing inter-aural mismatch with tonotopically matched FATs may benefit bilateral CI users' ability to segregate competing speech despite substantial low-frequency information loss in ears with shallow insertion depths.
Collapse
Affiliation(s)
- Mathew Thomas
- Department of Head and Neck Surgery, David Geffen School of Medicine, UCLA, Los Angeles, CA, United States of America
| | - Shelby Willis
- Department of Head and Neck Surgery, David Geffen School of Medicine, UCLA, Los Angeles, CA, United States of America
| | - John J. Galvin
- House Institute Foundation, Los Angeles, California, United States of America
| | - Qian-Jie Fu
- Department of Head and Neck Surgery, David Geffen School of Medicine, UCLA, Los Angeles, CA, United States of America
- * E-mail:
| |
Collapse
|
14
|
Xu K, Willis S, Gopen Q, Fu QJ. Effects of Spectral Resolution and Frequency Mismatch on Speech Understanding and Spatial Release From Masking in Simulated Bilateral Cochlear Implants. Ear Hear 2021; 41:1362-1371. [PMID: 32132377 DOI: 10.1097/aud.0000000000000865] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Due to interaural frequency mismatch, bilateral cochlear-implant (CI) users may be less able to take advantage of binaural cues that normal-hearing (NH) listeners use for spatial hearing, such as interaural time differences and interaural level differences. As such, bilateral CI users have difficulty segregating competing speech even when the target and competing talkers are spatially separated. The goal of this study was to evaluate the effects of spectral resolution, tonotopic mismatch (the frequency mismatch between the acoustic center frequency assigned to CI electrode within an implanted ear relative to the expected spiral ganglion characteristic frequency), and interaural mismatch (differences in the degree of tonotopic mismatch in each ear) on speech understanding and spatial release from masking (SRM) in the presence of competing talkers in NH subjects listening to bilateral vocoder simulations. DESIGN During testing, both target and masker speech were presented in five-word sentences that had the same syntax but were not necessarily meaningful. The sentences were composed of five categories in fixed order (Name, Verb, Number, Color, and Clothes), each of which had 10 items, such that multiple sentences could be generated by randomly selecting a word from each category. Speech reception thresholds (SRTs) for the target sentence presented in competing speech maskers were measured. The target speech was delivered to both ears and the two speech maskers were delivered to (1) both ears (diotic masker), or (2) different ears (dichotic masker: one delivered to the left ear and the other delivered to the right ear). Stimuli included the unprocessed speech and four 16-channel sine-vocoder simulations with different interaural mismatch (0, 1, and 2 mm). SRM was calculated as the difference between the diotic and dichotic listening conditions. RESULTS With unprocessed speech, SRTs were 0.3 and -18.0 dB for the diotic and dichotic maskers, respectively. For the spectrally degraded speech with mild tonotopic mismatch and no interaural mismatch, SRTs were 5.6 and -2.0 dB for the diotic and dichotic maskers, respectively. When the tonotopic mismatch increased in both ears, SRTs worsened to 8.9 and 2.4 dB for the diotic and dichotic maskers, respectively. When the two ears had different tonotopic mismatch (e.g., there was interaural mismatch), the performance drop in SRTs was much larger for the dichotic than for the diotic masker. The largest SRM was observed with unprocessed speech (18.3 dB). With the CI simulations, SRM was significantly reduced to 7.6 dB even with mild tonotopic mismatch but no interaural mismatch; SRM was further reduced with increasing interaural mismatch. CONCLUSIONS The results demonstrate that frequency resolution, tonotopic mismatch, and interaural mismatch have differential effects on speech understanding and SRM in simulation of bilateral CIs. Minimizing interaural mismatch may be critical to optimize binaural benefits and improve CI performance for competing speech, a typical listening environment. SRM (the difference in SRTs between diotic and dichotic maskers) may be a useful clinical tool to assess interaural frequency mismatch in bilateral CI users and to evaluate the benefits of optimization methods that minimize interaural mismatch.
Collapse
Affiliation(s)
- Kevin Xu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, California, USA
| | | | | | | |
Collapse
|
15
|
The smaller the frequency-to-place mismatch the better the hearing outcomes in cochlear implant recipients? Eur Arch Otorhinolaryngol 2021; 279:1875-1883. [PMID: 34131770 DOI: 10.1007/s00405-021-06899-y] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2020] [Accepted: 05/20/2021] [Indexed: 10/21/2022]
Abstract
OBJECTIVE To investigate the effect of frequency-to-place mismatch, i.e. the mismatch between the tonotopic frequency map in the cochlea and the frequency band that is assigned to an electrode contact of a cochlear implant (CI) at the same cochlear location on speech perception outcomes, using postoperative CT images. STUDY DESIGN Retrospective observational single-centre study. METHODS Retrospective pre- and postoperative clinical CT data of 39 CI recipients with normal cochlear anatomy were analysed in an otological surgical planning software. The tonotopic frequency at each electrode position was estimated using the Greenwood function. For each patient, frequency-to-place mismatch between the tonotopic frequency and the fitted centre frequency for each electrode contact was calculated. The influence of frequency-to-place mismatch on speech perception in noise at 6 and 12 months after CI activation was studied. RESULTS A significant linear correlation was found between the frequency-to-place mismatch and speech perception in noise 6 months after cochlear implantation (p < 0.05). The smaller the frequency-to-place mismatch, the better the initial speech perception in noise results of the CI recipients. The significant effect disappeared after 12 months CI experience. CONCLUSION The study findings support the idea of minimizing the frequency-to-place mismatch in CI recipients in order to pursue better initial speech perception in noise. Further research is needed to investigate the prospect of tonotopic fitting strategies based upon postoperative CT images of the exact locations of the electrode contacts.
Collapse
|
16
|
Individual Variability in Recalibrating to Spectrally Shifted Speech: Implications for Cochlear Implants. Ear Hear 2021; 42:1412-1427. [PMID: 33795617 DOI: 10.1097/aud.0000000000001043] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Cochlear implant (CI) recipients are at a severe disadvantage compared with normal-hearing listeners in distinguishing consonants that differ by place of articulation because the key relevant spectral differences are degraded by the implant. One component of that degradation is the upward shifting of spectral energy that occurs with a shallow insertion depth of a CI. The present study aimed to systematically measure the effects of spectral shifting on word recognition and phoneme categorization by specifically controlling the amount of shifting and using stimuli whose identification specifically depends on perceiving frequency cues. We hypothesized that listeners would be biased toward perceiving phonemes that contain higher-frequency components because of the upward frequency shift and that intelligibility would decrease as spectral shifting increased. DESIGN Normal-hearing listeners (n = 15) heard sine wave-vocoded speech with simulated upward frequency shifts of 0, 2, 4, and 6 mm of cochlear space to simulate shallow CI insertion depth. Stimuli included monosyllabic words and /b/-/d/ and /∫/-/s/ continua that varied systematically by formant frequency transitions or frication noise spectral peaks, respectively. Recalibration to spectral shifting was operationally defined as shifting perceptual acoustic-phonetic mapping commensurate with the spectral shift. In other words, adjusting frequency expectations for both phonemes upward so that there is still a perceptual distinction, rather than hearing all upward-shifted phonemes as the higher-frequency member of the pair. RESULTS For moderate amounts of spectral shifting, group data suggested a general "halfway" recalibration to spectral shifting, but individual data suggested a notably different conclusion: half of the listeners were able to recalibrate fully, while the other halves of the listeners were utterly unable to categorize shifted speech with any reliability. There were no participants who demonstrated a pattern intermediate to these two extremes. Intelligibility of words decreased with greater amounts of spectral shifting, also showing loose clusters of better- and poorer-performing listeners. Phonetic analysis of word errors revealed certain cues were more susceptible to being compromised due to a frequency shift (place and manner of articulation), while voicing was robust to spectral shifting. CONCLUSIONS Shifting the frequency spectrum of speech has systematic effects that are in line with known properties of speech acoustics, but the ensuing difficulties cannot be predicted based on tonotopic mismatch alone. Difficulties are subject to substantial individual differences in the capacity to adjust acoustic-phonetic mapping. These results help to explain why speech recognition in CI listeners cannot be fully predicted by peripheral factors like electrode placement and spectral resolution; even among listeners with functionally equivalent auditory input, there is an additional factor of simply being able or unable to flexibly adjust acoustic-phonetic mapping. This individual variability could motivate precise treatment approaches guided by an individual's relative reliance on wideband frequency representation (even if it is mismatched) or limited frequency coverage whose tonotopy is preserved.
Collapse
|
17
|
Newman RS, Morini G, Shroads E, Chatterjee M. Toddlers' fast-mapping from noise-vocoded speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:2432. [PMID: 32359241 PMCID: PMC7176458 DOI: 10.1121/10.0001129] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/30/2019] [Revised: 04/02/2020] [Accepted: 04/03/2020] [Indexed: 06/11/2023]
Abstract
The ability to recognize speech that is degraded spectrally is a critical skill for successfully using a cochlear implant (CI). Previous research has shown that toddlers with normal hearing can successfully recognize noise-vocoded words as long as the signal contains at least eight spectral channels [Newman and Chatterjee. (2013). J. Acoust. Soc. Am. 133(1), 483-494; Newman, Chatterjee, Morini, and Remez. (2015). J. Acoust. Soc. Am. 138(3), EL311-EL317], although they have difficulty with signals that only contain four channels of information. Young children with CIs not only need to match a degraded speech signal to a stored representation (word recognition), but they also need to create new representations (word learning), a task that is likely to be more cognitively demanding. Normal-hearing toddlers aged 34 months were tested on their ability to initially learn (fast-map) new words in noise-vocoded stimuli. While children were successful at fast-mapping new words from 16-channel noise-vocoded stimuli, they failed to do so from 8-channel noise-vocoded speech. The level of degradation imposed by 8-channel vocoding appears sufficient to disrupt fast-mapping in young children. Recent results indicate that only CI patients with high spectral resolution can benefit from more than eight active electrodes. This suggests that for many children with CIs, reduced spectral resolution may limit their acquisition of novel words.
Collapse
Affiliation(s)
- Rochelle S Newman
- Department of Hearing and Speech Sciences, University of Maryland, 0100 Lefrak Hall, College Park, Maryland 20742, USA
| | - Giovanna Morini
- Department of Communication Sciences and Disorders, University of Delaware, 100 Discovery Boulevard, Newark, Delaware 19713, USA
| | - Emily Shroads
- Department of Hearing and Speech Sciences, University of Maryland, 0100 Lefrak Hall, College Park, Maryland 20742, USA
| | - Monita Chatterjee
- Boys Town National Research Hospital, 555 North 30th Street, Omaha, Nebraska 68131, USA
| |
Collapse
|
18
|
Luo X, Garrett C. Dynamic current steering with phantom electrode in cochlear implants. Hear Res 2020; 390:107949. [PMID: 32200300 DOI: 10.1016/j.heares.2020.107949] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/31/2019] [Revised: 03/03/2020] [Accepted: 03/10/2020] [Indexed: 12/01/2022]
Abstract
Phantom electrode (PE) stimulation can extend the lower limit of pitch perception with cochlear implants (CIs) by using simultaneous out-of-phase stimulation of the most apical primary electrode and the adjacent basal compensating electrode. The total electrical field may push the excitation pattern beyond the most apical electrode to elicit a lower pitch, depending on the ratio of current between the compensating and primary electrodes (i.e., the compensation coefficient σ). This study tested the hypothesis that dynamic current steering of PE stimuli can be implemented by varying σ over time to encode spectral details in low frequencies. To determine the range of σ for current steering and the corresponding current levels, Experiment 1 tested CI users' loudness balance and pitch ranking of static PE stimuli with σ from 0 to 0.6 in steps of 0.2. It was found that the equal-loudness most comfortable level significantly increased with σ and can be modeled by a piecewise linear function of σ. Consistent with the previous findings, higher σ elicited either lower or similar pitches without salient pitch reversals than lower σ. Based on the results of Experiment 1, Experiment 2 created flat, rising, and falling pitch contours of 300-1000 ms using dynamic PE stimuli with time-varying σ from 0 to 0.6 and equal-loudness current levels. In a pitch contour identification (PCI) task, CI users scored 80% and above on average. Increasing the stimulus duration from 300 to 1000 ms slightly but did not significantly improve the PCI scores. Across subjects, the 1000-ms PCI scores in Experiment 2 were significantly correlated with the cumulative pitch-ranking sensitivity in Experiment 1. It is thus feasible to use dynamic current steering with PE to encode low-frequency pitch cues for CI users.
Collapse
Affiliation(s)
- Xin Luo
- Program of Speech and Hearing Science, College of Health Solutions, Arizona State University, 975 S. Myrtle Av., P.O. Box 870102, Tempe, AZ, 85287, USA.
| | - Christopher Garrett
- Program of Speech and Hearing Science, College of Health Solutions, Arizona State University, 975 S. Myrtle Av., P.O. Box 870102, Tempe, AZ, 85287, USA
| |
Collapse
|
19
|
Cardin V, Rosen S, Konieczny L, Coulson K, Lametti D, Edwards M, Woll B. The effect of dopamine on the comprehension of spectrally-shifted noise-vocoded speech: a pilot study. Int J Audiol 2020; 59:674-681. [PMID: 32186216 DOI: 10.1080/14992027.2020.1734675] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
Objectives: Cochlear implantation has proven beneficial in restoring hearing. However, success is variable, and there is a need for a simple post-implantation therapy that could significantly increase implantation success. Dopamine has a general role in learning and in assigning value to environmental stimuli. We tested the effect of dopamine in the comprehension of spectrally-shifted noise-vocoded (SSNV) speech, which simulates, in hearing individuals, the signal delivered by a cochlear implant (CI).Design and study sample: Thirty-five participants (age = 38.0 ± 10.1 SD) recruited from the general population were divided into three groups. We tested SSNV speech comprehension in two experimental sessions. In one session, a metabolic precursor of dopamine (L-DOPA) was administered to participants in two of the groups; a placebo was administered in the other session.Results: A single dose of L-DOPA interacted with training to improve perception of SSNV speech, but did not significantly accelerate learning.Conclusions: These findings are a first step in exploring the use of dopamine to enhance speech understanding in CI patients. Replications of these results using SSNV in individuals with normal hearing, and also in CI users, are needed to determine whether these effects can translate into benefits in everyday language comprehension.
Collapse
Affiliation(s)
- Velia Cardin
- Deafness, Cognition and Language Research Centre, University College London, London, United Kingdom.,School of Psychology, University of East Anglia, Norwich, Norfolk, United Kingdom
| | - Stuart Rosen
- Speech, Hearing and Phonetics Sciences, UCL, London, United Kingdom
| | - Linda Konieczny
- Deafness, Cognition and Language Research Centre, University College London, London, United Kingdom
| | - Kim Coulson
- Deafness, Cognition and Language Research Centre, University College London, London, United Kingdom
| | - Daniel Lametti
- Department of Psychology, Acadia University, Wolfville, Nova Scotia, Canada
| | - Mark Edwards
- Neuroscience Research Centre, Institute of Molecular and Clinical Sciences, St George's University of London, London, United Kingdom
| | - Bencie Woll
- Deafness, Cognition and Language Research Centre, University College London, London, United Kingdom
| |
Collapse
|
20
|
Karoui C, James C, Barone P, Bakhos D, Marx M, Macherey O. Searching for the Sound of a Cochlear Implant: Evaluation of Different Vocoder Parameters by Cochlear Implant Users With Single-Sided Deafness. Trends Hear 2020; 23:2331216519866029. [PMID: 31533581 PMCID: PMC6753516 DOI: 10.1177/2331216519866029] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Cochlear implantation in subjects with single-sided deafness (SSD) offers a unique opportunity to directly compare the percepts evoked by a cochlear implant (CI) with those evoked acoustically. Here, nine SSD-CI users performed a forced-choice task evaluating the similarity of speech processed by their CI with speech processed by several vocoders presented to their healthy ear. In each trial, subjects heard two intervals: their CI followed by a certain vocoder in Interval 1 and their CI followed by a different vocoder in Interval 2. The vocoders differed either (i) in carrier type-(sinusoidal [SINE], bandfiltered noise [NOISE], and pulse-spreading harmonic complex) or (ii) in frequency mismatch between the analysis and synthesis frequency ranges-(no mismatch, and two frequency-mismatched conditions of 2 and 4 equivalent rectangular bandwidths [ERBs]). Subjects had to state in which of the two intervals the CI and vocoder sounds were more similar. Despite a large intersubject variability, the PSHC vocoder was judged significantly more similar to the CI than SINE or NOISE vocoders. Furthermore, the No-mismatch and 2-ERB mismatch vocoders were judged significantly more similar to the CI than the 4-ERB mismatch vocoder. The mismatch data were also interpreted by comparing spiral ganglion characteristic frequencies with electrode contact positions determined from postoperative computed tomography scans. Only one subject demonstrated a pattern of preference consistent with adaptation to the CI sound processor frequency-to-electrode allocation table and two subjects showed possible partial adaptation. Those subjects with adaptation patterns presented overall small and consistent frequency mismatches across their electrode arrays.
Collapse
Affiliation(s)
- Chadlia Karoui
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Cochlear France SAS, Toulouse, France
| | - Chris James
- Cochlear France SAS, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Pascal Barone
- Centre de Recherche Cerveau et Cognition, Toulouse, France
| | - David Bakhos
- Université François-Rabelais de Tours, CHRU de Tours, France.,Ear Nose and Throat department, CHUR de Tours, Tours, France
| | - Mathieu Marx
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Olivier Macherey
- Aix Marseille University, CNRS, Centrale Marseille, LMA, Marseille, France
| |
Collapse
|
21
|
Sadeghi M, Zhai X, Stevenson IH, Escabí MA. A neural ensemble correlation code for sound category identification. PLoS Biol 2019; 17:e3000449. [PMID: 31574079 PMCID: PMC6788721 DOI: 10.1371/journal.pbio.3000449] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2019] [Revised: 10/11/2019] [Accepted: 09/03/2019] [Indexed: 12/25/2022] Open
Abstract
Humans and other animals effortlessly identify natural sounds and categorize them into behaviorally relevant categories. Yet, the acoustic features and neural transformations that enable sound recognition and the formation of perceptual categories are largely unknown. Here, using multichannel neural recordings in the auditory midbrain of unanesthetized female rabbits, we first demonstrate that neural ensemble activity in the auditory midbrain displays highly structured correlations that vary with distinct natural sound stimuli. These stimulus-driven correlations can be used to accurately identify individual sounds using single-response trials, even when the sounds do not differ in their spectral content. Combining neural recordings and an auditory model, we then show how correlations between frequency-organized auditory channels can contribute to discrimination of not just individual sounds but sound categories. For both the model and neural data, spectral and temporal correlations achieved similar categorization performance and appear to contribute equally. Moreover, both the neural and model classifiers achieve their best task performance when they accumulate evidence over a time frame of approximately 1-2 seconds, mirroring human perceptual trends. These results together suggest that time-frequency correlations in sounds may be reflected in the correlations between auditory midbrain ensembles and that these correlations may play an important role in the identification and categorization of natural sounds.
Collapse
Affiliation(s)
- Mina Sadeghi
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Xiu Zhai
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
| | - Ian H. Stevenson
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
| | - Monty A. Escabí
- Department of Electrical and Computer Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Biomedical Engineering, University of Connecticut, Storrs, Connecticut, United States of America
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, United States of America
- * E-mail:
| |
Collapse
|
22
|
The Effect of Simulated Interaural Frequency Mismatch on Speech Understanding and Spatial Release From Masking. Ear Hear 2019; 39:895-905. [PMID: 29337763 DOI: 10.1097/aud.0000000000000541] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVE The binaural-hearing system interaurally compares inputs, which underlies the ability to localize sound sources and to better understand speech in complex acoustic environments. Cochlear implants (CIs) are provided in both ears to increase binaural-hearing benefits; however, bilateral CI users continue to struggle with understanding speech in the presence of interfering sounds and do not achieve the same level of spatial release from masking (SRM) as normal-hearing listeners. One reason for diminished SRM in CI users could be that the electrode arrays are inserted at different depths in each ear, which would cause an interaural frequency mismatch. Because interaural frequency mismatch diminishes the salience of interaural differences for relatively simple stimuli, it may also diminish binaural benefits for spectral-temporally complex stimuli like speech. This study evaluated the effect of simulated frequency-to-place mismatch on speech understanding and SRM. DESIGN Eleven normal-hearing listeners were tested on a speech understanding task. There was a female target talker who spoke five-word sentences from a closed set of words. There were two interfering male talkers who spoke unrelated sentences. Nonindividualized head-related transfer functions were used to simulate a virtual auditory space. The target was presented from the front (0°), and the interfering speech was either presented from the front (colocated) or from 90° to the right (spatially separated). Stimuli were then processed by an eight-channel vocoder with tonal carriers to simulate aspects of listening through a CI. Frequency-to-place mismatch ("shift") was introduced by increasing the center frequency of the synthesis filters compared with the corresponding analysis filters. Speech understanding was measured for different shifts (0, 3, 4.5, and 6 mm) and target-to-masker ratios (TMRs: +10 to -10 dB). SRM was calculated as the difference in the percentage of correct words for the colocated and separated conditions. Two types of shifts were tested: (1) bilateral shifts that had the same frequency-to-place mismatch in both ears, but no interaural frequency mismatch, and (2) unilateral shifts that produced an interaural frequency mismatch. RESULTS For the bilateral shift conditions, speech understanding decreased with increasing shift and with decreasing TMR, for both colocated and separate conditions. There was, however, no interaction between shift and spatial configuration; in other words, SRM was not affected by shift. For the unilateral shift conditions, speech understanding decreased with increasing interaural mismatch and with decreasing TMR for both the colocated and spatially separated conditions. Critically, there was a significant interaction between the amount of shift and spatial configuration; in other words, SRM decreased for increasing interaural mismatch. CONCLUSIONS A frequency-to-place mismatch in one or both ears resulted in decreased speech understanding. SRM, however, was only affected in conditions with unilateral shifts and interaural frequency mismatch. Therefore, matching frequency information between the ears provides listeners with larger binaural-hearing benefits, for example, improved speech understanding in the presence of interfering talkers. A clinical procedure to reduce interaural frequency mismatch when programming bilateral CIs may improve benefits in speech segregation that are due to binaural-hearing abilities.
Collapse
|
23
|
Grasmeder ML, Verschuur CA, van Besouw RM, Wheatley AMH, Newman TA. Measurement of pitch perception as a function of cochlear implant electrode and its effect on speech perception with different frequency allocations. Int J Audiol 2018; 58:158-166. [PMID: 30370800 DOI: 10.1080/14992027.2018.1516048] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/28/2022]
Abstract
OBJECTIVE An experiment was conducted to investigate the possibility that speech perception could be improved for some cochlear implant (CI) users by adjustment of the frequency allocation to the electrodes, following assessment of pitch perception along the electrode array. STUDY SAMPLE Thirteen adult CI users with MED-EL devices participated in the study. DESIGN Pitch perception was assessed for individual CI electrode pairs using the Pitch Contour Test (PCT), giving information on pitch discrimination and pitch ranking for adjacent electrodes. Sentence perception in noise was also assessed with ten different frequency allocations, including the default. RESULTS Pitch perception was found to be poorer for both discrimination and ranking scores at either end of the electrode array. A significant effect of frequency allocation was found for sentence scores [F(4.24,38.2) = 7.14, p < 0.001] and a significant interaction between sentence score and PCT ranking score for basal electrodes was found [F(4.24,38.2) = 2.95, p = 0.03]. Participants with poorer pitch perception at the basal end had poorer scores for some allocations with greater basal shift. CONCLUSIONS The results suggest that speech perception could be improved for CI users by assessment of pitch perception using the PCT and subsequent adjustment of pitch-related stimulation parameters.
Collapse
Affiliation(s)
- M L Grasmeder
- a Auditory Implant Service University of Southampton , Southampton , UK
| | - C A Verschuur
- a Auditory Implant Service University of Southampton , Southampton , UK
| | - R M van Besouw
- b Institute of Sound and Vibration Research, University of Southampton , UK
| | - A M H Wheatley
- b Institute of Sound and Vibration Research, University of Southampton , UK
| | - T A Newman
- c Southampton Neuroscience Group , University of Southampton , UK
| |
Collapse
|
24
|
Assessing Cochlear Length Using Cone Beam Computed Tomography in Adults With Cochlear Implants. Otol Neurotol 2018; 39:e757-e764. [DOI: 10.1097/mao.0000000000001934] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/26/2022]
|
25
|
El Boghdady N, Başkent D, Gaudrain E. Effect of frequency mismatch and band partitioning on vocal tract length perception in vocoder simulations of cochlear implant processing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:3505. [PMID: 29960490 DOI: 10.1121/1.5041261] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
The vocal tract length (VTL) of a speaker is an important voice cue that aids speech intelligibility in multi-talker situations. However, cochlear implant (CI) users demonstrate poor VTL sensitivity. This may be partially caused by the mismatch between frequencies received by the implant and those corresponding to places of stimulation along the cochlea. This mismatch can distort formant spacing, where VTL cues are encoded. In this study, the effects of frequency mismatch and band partitioning on VTL sensitivity were investigated in normal hearing listeners with vocoder simulations of CI processing. The hypotheses were that VTL sensitivity may be reduced by increased frequency mismatch and insufficient spectral resolution in how the frequency range is partitioned, specifically where formants lie. Moreover, optimal band partitioning might mitigate the detrimental effects of frequency mismatch on VTL sensitivity. Results showed that VTL sensitivity decreased with increased frequency mismatch and reduced spectral resolution near the low frequencies of the band partitioning map. Band partitioning was independent of mismatch, indicating that if a given partitioning is suboptimal, a better partitioning might improve VTL sensitivity despite the degree of mismatch. These findings suggest that customizing the frequency partitioning map may enhance VTL perception in individual CI users.
Collapse
Affiliation(s)
- Nawal El Boghdady
- University of Groningen, University Medical Center Groningen, Department of Otorhinolaryngology/Head and Neck Surgery, Groningen, The Netherlands
| | - Deniz Başkent
- University of Groningen, University Medical Center Groningen, Department of Otorhinolaryngology/Head and Neck Surgery, Groningen, The Netherlands
| | - Etienne Gaudrain
- University of Groningen, University Medical Center Groningen, Department of Otorhinolaryngology/Head and Neck Surgery, Groningen, The Netherlands
| |
Collapse
|
26
|
Mistrík P, Jolly C, Sieber D, Hochmair I. Challenging aspects of contemporary cochlear implant electrode array design. World J Otorhinolaryngol Head Neck Surg 2018; 3:192-199. [PMID: 29780962 PMCID: PMC5956130 DOI: 10.1016/j.wjorl.2017.12.007] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2017] [Accepted: 12/12/2017] [Indexed: 01/25/2023] Open
Abstract
Objective A design comparison of current perimodiolar and lateral wall electrode arrays of the cochlear implant (CI) is provided. The focus is on functional features such as acoustic frequency coverage and tonotopic mapping, battery consumption and dynamic range. A traumacity of their insertion is also evaluated. Methods Review of up-to-date literature. Results Perimodiolar electrode arrays are positioned in the basal turn of the cochlea near the modiolus. They are designed to initiate the action potential in the proximity to the neural soma located in spiral ganglion. On the other hand, lateral wall electrode arrays can be inserted deeper inside the cochlea, as they are located along the lateral wall and such insertion trajectory is less traumatic. This class of arrays targets primarily surviving neural peripheral processes. Due to their larger insertion depth, lateral wall arrays can deliver lower acoustic frequencies in manner better corresponding to cochlear tonotopicity. In fact, spiral ganglion sections containing auditory nerve fibres tuned to low acoustic frequencies are located deeper than 1 and half turn inside the cochlea. For this reason, a significant frequency mismatch might be occurring for apical electrodes in perimodiolar arrays, detrimental to speech perception. Tonal languages such as Mandarin might be therefore better treated with lateral wall arrays. On the other hand, closer proximity to target tissue results in lower psychophysical threshold levels for perimodiolar arrays. However, the maximal comfort level is also lower, paradoxically resulting in narrower dynamic range than that of lateral wall arrays. Battery consumption is comparable for both types of arrays. Conclusions Lateral wall arrays are less likely to cause trauma to cochlear structures. As the current trend in cochlear implantation is the maximal protection of residual acoustic hearing, the lateral wall arrays seem more suitable for hearing preservation CI surgeries. Future development could focus on combining the advantages of both types: perimodiolar location in the basal turn extended to lateral wall location for higher turn locations.
Collapse
|
27
|
Abstract
OBJECTIVE The goal of the present study was to assess the sound quality of a cochlear implant for single-sided deaf (SSD) patients fit with a cochlear implant (CI). BACKGROUND One of the fundamental, unanswered questions in CI research is "what does an implant sound like?" Conventional CI patients must use the memory of a clean signal, often decades old, to judge the sound quality of their CIs. In contrast, SSD-CI patients can rate the similarity of a clean signal presented to the CI ear and candidate, CI-like signals presented to the ear with normal hearing. METHODS For Experiment 1 four types of stimuli were created for presentation to the normal hearing ear: noise vocoded signals, sine vocoded signals, frequency shifted, sine vocoded signals and band-pass filtered, natural speech signals. Listeners rated the similarity of these signals to unmodified signals sent to the CI on a scale of 0 to 10 with 10 being a complete match to the CI signal. For Experiment 2 multitrack signal mixing was used to create natural speech signals that varied along multiple dimensions. RESULTS In Experiment 1 for eight adult SSD-CI listeners, the best median similarity rating to the sound of the CI for noise vocoded signals was 1.9; for sine vocoded signals 2.9; for frequency upshifted signals, 1.9; and for band pass filtered signals, 5.5. In Experiment 2 for three young listeners, combinations of band pass filtering and spectral smearing lead to ratings of 10. CONCLUSION The sound quality of noise and sine vocoders does not generally correspond to the sound quality of cochlear implants fit to SSD patients. Our preliminary conclusion is that natural speech signals that have been muffled to one degree or another by band pass filtering and/or spectral smearing provide a close, but incomplete, match to CI sound quality for some patients.
Collapse
|
28
|
An SY, An CH, Lee KY, Jang JH, Choung YH, Lee SH. Diagnostic role of cone beam computed tomography for the position of straight array. Acta Otolaryngol 2018; 138:375-381. [PMID: 29172857 DOI: 10.1080/00016489.2017.1404639] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Abstract
OBJECTIVE To assess the usefulness of cone beam computed tomography (CBCT) for characterizing electrode insertion and evaluate the influence of electrode insertion status on post-cochlear implantation (CI) outcomes. DESIGN Twenty-six ears with post-CI CBCT scans were included. The devices were MED-EL Flex28 (n = 21) and Nucleus slim straight (n = 5). The parameters including cochlear duct length (CDL), insertion depth angle (IDA), insertion length of electrode (IL), and cochlear coverage (CC) were analyzed and compared with aided pure-tone threshold (PTA) with implant in free field, and open-set sentence score. RESULTS The mean CDL was 36.8 ± 1.4 mm. Electrode array was dislocated into scala tympani in two ears. The mean IL and IDA were 26.5 ± 1.9 mm and 541.4 ± 70.2°. The mean linear CC (IL/CDL, 0.73 ± 0.06) was larger than the mean angular CC (IDA/900, 0.60 ± 0.08). The CBCT parameters showed correlation one another. While the aided pure-tone threshold was correlated with IL and IDA, there were no significant correlations in the open-set sentence score. For the postlingually deaf patients with single electrode (Flex 28), the sentence score had no significant correlation and the aided PTA was positively correlated with IL (R = 0.517, p = .028). CONCLUSIONS This study validated the CBCT evaluating the electrode array position. The CBCT could be helpful for the preoperative selection of the optimal array and prediction of the CC.
Collapse
Affiliation(s)
- Seo-Young An
- Department of Oral and Maxillofacial Radiology, Kyungpook National University College of Dentistry, Daegu, Republic of Korea
| | - Chang-Hyeon An
- Department of Oral and Maxillofacial Radiology, Kyungpook National University College of Dentistry, Daegu, Republic of Korea
| | - Kyu-Yup Lee
- Department of Otorhinolaryngology, Kyungpook National University College of Medicine, Daegu, Republic of Korea
| | - Jeong Hun Jang
- Department of Otorhinolaryngology, Ajou University College of Medicine, Suwon, Republic of Korea
| | - Yun-Hoon Choung
- Department of Otorhinolaryngology, Ajou University College of Medicine, Suwon, Republic of Korea
| | - Sang Heun Lee
- Department of Otorhinolaryngology, Daegu Veterans Hospital, Daegu, Republic of Korea
| |
Collapse
|
29
|
Chang SA, Won JH, Kim H, Oh SH, Tyler RS, Cho CH. Frequency-Limiting Effects on Speech and Environmental Sound Identification for Cochlear Implant and Normal Hearing Listeners. J Audiol Otol 2018; 22:28-38. [PMID: 29325391 PMCID: PMC5784366 DOI: 10.7874/jao.2017.00178] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2017] [Revised: 10/09/2017] [Accepted: 10/17/2017] [Indexed: 11/22/2022] Open
Abstract
BACKGROUND AND OBJECTIVES It is important to understand the frequency region of cues used, and not used, by cochlear implant (CI) recipients. Speech and environmental sound recognition by individuals with CI and normal-hearing (NH) was measured. Gradients were also computed to evaluate the pattern of change in identification performance with respect to the low-pass filtering or high-pass filtering cutoff frequencies. SUBJECTS AND METHODS Frequency-limiting effects were implemented in the acoustic waveforms by passing the signals through low-pass filters (LPFs) or high-pass filters (HPFs) with seven different cutoff frequencies. Identification of Korean vowels and consonants produced by a male and female speaker and environmental sounds was measured. Crossover frequencies were determined for each identification test, where the LPF and HPF conditions show the identical identification scores. RESULTS CI and NH subjects showed changes in identification performance in a similar manner as a function of cutoff frequency for the LPF and HPF conditions, suggesting that the degraded spectral information in the acoustic signals may similarly constraint the identification performance for both subject groups. However, CI subjects were generally less efficient than NH subjects in using the limited spectral information for speech and environmental sound identification due to the inefficient coding of acoustic cues through the CI sound processors. CONCLUSIONS This finding will provide vital information in Korean for understanding how different the frequency information is in receiving speech and environmental sounds by CI processor from normal hearing.
Collapse
Affiliation(s)
- Son-A Chang
- Department of Otolaryngology-Head and Neck Surgery, Seoul National University Hospital, Seoul, Korea
| | - Jong Ho Won
- Department of Audiology and Speech Pathology, University of Tennessee Health Science Center, Knoxville, TN, USA
| | - HyangHee Kim
- Graduate Program of Speech and Language Pathology, Department and Research Institute of Rehabilitation Medicine, Yonsei University College of Medicine, Seoul, Korea
| | - Seung-Ha Oh
- Department of Otolaryngology-Head and Neck Surgery, Seoul National University Hospital, Seoul, Korea
| | - Richard S Tyler
- Department of Otolaryngology-Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA, USA
| | - Chang Hyun Cho
- Department of Otolaryngology-Head and Neck Surgery, Gachon University Gil Medical Center, Incheon, Korea
| |
Collapse
|
30
|
Integration of acoustic and electric hearing is better in the same ear than across ears. Sci Rep 2017; 7:12500. [PMID: 28970567 PMCID: PMC5624923 DOI: 10.1038/s41598-017-12298-3] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2017] [Accepted: 09/06/2017] [Indexed: 11/26/2022] Open
Abstract
Advances in cochlear implant (CI) technology allow for acoustic and electric hearing to be combined within the same ear (electric-acoustic stimulation, or EAS) and/or across ears (bimodal listening). Integration efficiency (IE; the ratio between observed and predicted performance for acoustic-electric hearing) can be used to estimate how well acoustic and electric hearing are combined. The goal of this study was to evaluate factors that affect IE in EAS and bimodal listening. Vowel recognition was measured in normal-hearing subjects listening to simulations of unimodal, EAS, and bimodal listening. The input/output frequency range for acoustic hearing was 0.1–0.6 kHz. For CI simulations, the output frequency range was 1.2–8.0 kHz to simulate a shallow insertion depth and the input frequency range was varied to provide increasing amounts of speech information and tonotopic mismatch. Performance was best when acoustic and electric hearing was combined in the same ear. IE was significantly better for EAS than for bimodal listening; IE was sensitive to tonotopic mismatch for EAS, but not for bimodal listening. These simulation results suggest acoustic and electric hearing may be more effectively and efficiently combined within rather than across ears, and that tonotopic mismatch should be minimized to maximize the benefit of acoustic-electric hearing, especially for EAS.
Collapse
|
31
|
Fitzgerald MB, Prosolovich K, Tan CT, Glassman EK, Svirsky MA. Self-Selection of Frequency Tables with Bilateral Mismatches in an Acoustic Simulation of a Cochlear Implant. J Am Acad Audiol 2017; 28:385-394. [PMID: 28534729 PMCID: PMC5563263 DOI: 10.3766/jaaa.15077] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
BACKGROUND Many recipients of bilateral cochlear implants (CIs) may have differences in electrode insertion depth. Previous reports indicate that when a bilateral mismatch is imposed, performance on tests of speech understanding or sound localization becomes worse. If recipients of bilateral CIs cannot adjust to a difference in insertion depth, adjustments to the frequency table may be necessary to maximize bilateral performance. PURPOSE The purpose of this study was to examine the feasibility of using real-time manipulations of the frequency table to offset any decrements in performance resulting from a bilateral mismatch. RESEARCH DESIGN A simulation of a CI was used because it allows for explicit control of the size of a bilateral mismatch. Such control is not available with users of CIs. STUDY SAMPLE A total of 31 normal-hearing young adults participated in this study. DATA COLLECTION AND ANALYSIS Using a CI simulation, four bilateral mismatch conditions (0, 0.75, 1.5, and 3 mm) were created. In the left ear, the analysis filters and noise bands of the CI simulation were the same. In the right ear, the noise bands were shifted higher in frequency to simulate a bilateral mismatch. Then, listeners selected a frequency table in the right ear that was perceived as maximizing bilateral speech intelligibility. Word-recognition scores were then assessed for each bilateral mismatch condition. Listeners were tested with both a standard frequency table, which preserved a bilateral mismatch, or with their self-selected frequency table. RESULTS Consistent with previous reports, bilateral mismatches of 1.5 and 3 mm yielded decrements in word recognition when the standard table was used in both ears. However, when listeners used the self-selected frequency table, performance was the same regardless of the size of the bilateral mismatch. CONCLUSIONS Self-selection of a frequency table appears to be a feasible method for ameliorating the negative effects of a bilateral mismatch. These data may have implications for recipients of bilateral CIs who cannot adapt to a bilateral mismatch, because they suggest that (1) such individuals may benefit from modification of the frequency table in one ear and (2) self-selection of a "most intelligible" frequency table may be a useful tool for determining how the frequency table should be altered to optimize speech recognition.
Collapse
Affiliation(s)
- Matthew B. Fitzgerald
- Department of Otolaryngology – Head and Neck Surgery, Stanford Ear Institute, Stanford University, Palo Alto, CA
- Department of Otolaryngology, New York University School of Medicine, New York, NY
| | - Ksenia Prosolovich
- Department of Otolaryngology, University of Southern California, Los Angeles, CA
| | - Chin-Tuan Tan
- Department of Otolaryngology, New York University School of Medicine, New York, NY
| | | | - Mario A. Svirsky
- Department of Otolaryngology, New York University School of Medicine, New York, NY
| |
Collapse
|
32
|
DiNino M, Wright RA, Winn MB, Bierer JA. Vowel and consonant confusions from spectrally manipulated stimuli designed to simulate poor cochlear implant electrode-neuron interfaces. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:4404. [PMID: 28039993 PMCID: PMC5392103 DOI: 10.1121/1.4971420] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/22/2016] [Revised: 10/15/2016] [Accepted: 11/22/2016] [Indexed: 05/26/2023]
Abstract
Suboptimal interfaces between cochlear implant (CI) electrodes and auditory neurons result in a loss or distortion of spectral information in specific frequency regions, which likely decreases CI users' speech identification performance. This study exploited speech acoustics to model regions of distorted CI frequency transmission to determine the perceptual consequences of suboptimal electrode-neuron interfaces. Normal hearing adults identified naturally spoken vowels and consonants after spectral information was manipulated through a noiseband vocoder: either (1) low-, middle-, or high-frequency regions of information were removed by zeroing the corresponding channel outputs, or (2) the same regions were distorted by splitting filter outputs to neighboring filters. These conditions simulated the detrimental effects of suboptimal CI electrode-neuron interfaces on spectral transmission. Vowel and consonant confusion patterns were analyzed with sequential information transmission, perceptual distance, and perceptual vowel space analyses. Results indicated that both types of spectral manipulation were equally destructive. Loss or distortion of frequency information produced similar effects on phoneme identification performance and confusion patterns. Consonant error patterns were consistently based on place of articulation. Vowel confusions showed that perceptions gravitated away from the degraded frequency region in a predictable manner, indicating that vowels can probe frequency-specific regions of spectral degradations.
Collapse
Affiliation(s)
- Mishaela DiNino
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd Street, Box 354875, Seattle, Washington 98105, USA
| | - Richard A Wright
- Department of Linguistics, University of Washington, Guggenheim Hall, Box 352425, Seattle, Washington, 98195, USA
| | - Matthew B Winn
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd Street, Box 354875, Seattle, Washington 98105, USA
| | - Julie Arenberg Bierer
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd Street, Box 354875, Seattle, Washington 98105, USA
| |
Collapse
|
33
|
Patro C, Mendel LL. Role of contextual cues on the perception of spectrally reduced interrupted speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2016; 140:1336. [PMID: 27586760 DOI: 10.1121/1.4961450] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
Understanding speech within an auditory scene is constantly challenged by interfering noise in suboptimal listening environments when noise hinders the continuity of the speech stream. In such instances, a typical auditory-cognitive system perceptually integrates available speech information and "fills in" missing information in the light of semantic context. However, individuals with cochlear implants (CIs) find it difficult and effortful to understand interrupted speech compared to their normal hearing counterparts. This inefficiency in perceptual integration of speech could be attributed to further degradations in the spectral-temporal domain imposed by CIs making it difficult to utilize the contextual evidence effectively. To address these issues, 20 normal hearing adults listened to speech that was spectrally reduced and spectrally reduced interrupted in a manner similar to CI processing. The Revised Speech Perception in Noise test, which includes contextually rich and contextually poor sentences, was used to evaluate the influence of semantic context on speech perception. Results indicated that listeners benefited more from semantic context when they listened to spectrally reduced speech alone. For the spectrally reduced interrupted speech, contextual information was not as helpful under significant spectral reductions, but became beneficial as the spectral resolution improved. These results suggest top-down processing facilitates speech perception up to a point, and it fails to facilitate speech understanding when the speech signals are significantly degraded.
Collapse
Affiliation(s)
- Chhayakanta Patro
- School of Communication Sciences and Disorders, University of Memphis, 4055 North Park Loop, Memphis, Tennessee, 38152, USA
| | - Lisa Lucks Mendel
- School of Communication Sciences and Disorders, University of Memphis, 4055 North Park Loop, Memphis, Tennessee, 38152, USA
| |
Collapse
|
34
|
Skinner MW, Holden TA, Whiting BR, Voie AH, Brunsden B, Neely JG, Saxon EA, Hullar TE, Finley CC. In Vivo Estimates of the Position of Advanced Bionics Electrode Arrays in the Human Cochlea. Ann Otol Rhinol Laryngol 2016. [DOI: 10.1177/00034894071160s401] [Citation(s) in RCA: 116] [Impact Index Per Article: 14.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Objectives: A new technique for determining the position of each electrode in the cochlea is described and applied to spiral computed tomography data from 15 patients implanted with Advanced Bionics HiFocus I, Ij, or Helix arrays. Methods: ANALYZE imaging software was used to register 3-dimensional image volumes from patients' preoperative and postoperative scans and from a single body donor whose unimplanted ears were scanned clinically, with micro computed tomography and with orthogonal-plane fluorescence optical sectioning (OPFOS) microscopy. By use of this registration, we compared the atlas of OPFOS images of soft tissue within the body donor's cochlea with the bone and fluid/tissue boundary available in patient scan data to choose the midmodiolar axis position and judge the electrode position in the scala tympani or scala vestibuli, including the distance to the medial and lateral scalar walls. The angular rotation 0° start point is a line joining the midmodiolar axis and the middle of the cochlear canal entry from the vestibule. Results: The group mean array insertion depth was 477° (range, 286° to 655°). The word scores were negatively correlated (r = −0.59; p = .028) with the number of electrodes in the scala vestibuli. Conclusions: Although the individual variability in all measures was large, repeated patterns of suboptimal electrode placement were observed across subjects, underscoring the applicability of this technique.
Collapse
|
35
|
Skinner MW, Holden TA, Whiting BR, Voie AH, Brunsden B, Neely JG, Saxon EA, Hullar TE, Finley CC. In Vivo Estimates of the Position of Advanced Bionics Electrode Arrays in the Human Cochlea. Ann Otol Rhinol Laryngol 2016. [DOI: 10.1177/000348940711600401] [Citation(s) in RCA: 48] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Objectives: A new technique for determining the position of each electrode in the cochlea is described and applied to spiral computed tomography data from 15 patients implanted with Advanced Bionics HiFocus I, Ij, or Helix arrays. Methods: ANALYZE imaging software was used to register 3-dimensional image volumes from patients' preoperative and postoperative scans and from a single body donor whose unimplanted ears were scanned clinically, with micro computed tomography and with orthogonal-plane fluorescence optical sectioning (OPFOS) microscopy. By use of this registration, we compared the atlas of OPFOS images of soft tissue within the body donor's cochlea with the bone and fluid/ tissue boundary available in patient scan data to choose the midmodiolar axis position and judge the electrode position in the scala tympani or scala vestibuli, including the distance to the medial and lateral scalar walls. The angular rotation 0° start point is a line joining the midmodiolar axis and the middle of the cochlear canal entry from the vestibule. Results: The group mean array insertion depth was 477° (range, 286° to 655°). The word scores were negatively correlated (r = −0.59; p = .028) with the number of electrodes in the scala vestibuli. Conclusions: Although the individual variability in all measures was large, repeated patterns of suboptimal electrode placement were observed across subjects, underscoring the applicability of this technique.
Collapse
|
36
|
Landsberger DM, Svrakic M, Roland JT, Svirsky M. The Relationship Between Insertion Angles, Default Frequency Allocations, and Spiral Ganglion Place Pitch in Cochlear Implants. Ear Hear 2016; 36:e207-13. [PMID: 25860624 DOI: 10.1097/aud.0000000000000163] [Citation(s) in RCA: 172] [Impact Index Per Article: 21.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Commercially available cochlear implant systems attempt to deliver frequency information going down to a few hundred Hertz, but the electrode arrays are not designed to reach the most apical regions of the cochlea, which correspond to these low frequencies. This may cause a mismatch between the frequencies presented by a cochlear implant electrode array and the frequencies represented at the corresponding location in a normal-hearing cochlea. In the following study, the mismatch between the frequency presented at a given cochlear angle and the frequency expected by an acoustic hearing ear at the corresponding angle is examined for the cochlear implant systems that are most commonly used in the United States. DESIGN The angular insertion of each of the electrodes on four different electrode arrays (MED-EL Standard, MED-EL Flex28, Advanced Bionics HiFocus 1J, and Cochlear Contour Advance) was estimated from X-ray. For the angular location of each electrode on each electrode array, the predicted spiral ganglion frequency was estimated. The predicted spiral ganglion frequency was compared with the center frequency provided by the corresponding electrode using the manufacturer's default frequency-to-electrode allocation. RESULTS Differences across devices were observed for the place of stimulation for frequencies below 650 Hz. Longer electrode arrays (i.e., the MED-EL Standard and Flex28) demonstrated smaller deviations from the spiral ganglion map than the other electrode arrays. For insertion angles up to approximately 270°, the frequencies presented at a given location were typically approximately an octave below what would be expected by a spiral ganglion frequency map, while the deviations were larger for angles deeper than 270°. For frequencies above 650 Hz, the frequency to angle relationship was consistent across all four electrode models. CONCLUSIONS A mismatch was observed between the predicted frequency and the default frequency provided by every electrode on all electrode arrays. The mismatch can be reduced by changing the default frequency allocations, inserting electrodes deeper into the cochlea, or allowing cochlear implant users to adapt to the mismatch. Further studies are required to fully assess the clinical significance of the frequency mismatch.
Collapse
Affiliation(s)
- David M Landsberger
- Department of Otolaryngology, New York University School of Medicine, New York, USA
| | | | | | | |
Collapse
|
37
|
Mistrík P, Jolly C. Optimal electrode length to match patient specific cochlear anatomy. Eur Ann Otorhinolaryngol Head Neck Dis 2016; 133 Suppl 1:S68-71. [DOI: 10.1016/j.anorl.2016.05.001] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2016] [Accepted: 05/03/2016] [Indexed: 11/15/2022]
|
38
|
Abstract
Music and speech share many acoustic cues but not all are equally important. For example, harmonic pitch is essential for music but not for speech. When birds communicate is their song more like speech or music? A new study contrasting pitch and spectral patterns shows that birds perceive their song more like humans perceive speech.
Collapse
|
39
|
El Boghdady N, Kegel A, Lai WK, Dillier N. A neural-based vocoder implementation for evaluating cochlear implant coding strategies. Hear Res 2016; 333:136-149. [PMID: 26775182 DOI: 10.1016/j.heares.2016.01.005] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/23/2015] [Revised: 12/18/2015] [Accepted: 01/07/2016] [Indexed: 10/22/2022]
Abstract
Most simulations of cochlear implant (CI) coding strategies rely on standard vocoders that are based on purely signal processing techniques. However, these models neither account for various biophysical phenomena, such as neural stochasticity and refractoriness, nor for effects of electrical stimulation, such as spectral smearing as a function of stimulus intensity. In this paper, a neural model that accounts for stochastic firing, parasitic spread of excitation across neuron populations, and neuronal refractoriness, was developed and augmented as a preprocessing stage for a standard 22-channel noise-band vocoder. This model was used to subjectively and objectively assess consonant discrimination in commercial and experimental coding strategies. Stimuli consisting of consonant-vowel (CV) and vowel-consonant-vowel (VCV) tokens were processed by either the Advanced Combination Encoder (ACE) or the Excitability Controlled Coding (ECC) strategies, and later resynthesized to audio using the aforementioned vocoder model. Baseline performance was measured using unprocessed versions of the speech tokens. Behavioural responses were collected from seven normal hearing (NH) volunteers, while EEG data were recorded from five NH participants. Psychophysical results indicate that while there may be a difference in consonant perception between the two tested coding strategies, mismatch negativity (MMN) waveforms do not show any marked trends in CV or VCV contrast discrimination.
Collapse
Affiliation(s)
- Nawal El Boghdady
- Institute for Neuroinformatics (INI), Universität Zürich (UZH)/ ETH Zürich (ETHZ), Zürich, Switzerland.
| | - Andrea Kegel
- Laboratory of Experimental Audiology, ENT Department, Universitätsspital Zürich (USZ), Zürich, Switzerland
| | - Wai Kong Lai
- Laboratory of Experimental Audiology, ENT Department, Universitätsspital Zürich (USZ), Zürich, Switzerland
| | - Norbert Dillier
- Laboratory of Experimental Audiology, ENT Department, Universitätsspital Zürich (USZ), Zürich, Switzerland
| |
Collapse
|
40
|
Dillon MT, Buss E, Adunka OF, Buchman CA, Pillsbury HC. Influence of Test Condition on Speech Perception With Electric-Acoustic Stimulation. Am J Audiol 2015; 24:520-8. [PMID: 26650652 DOI: 10.1044/2015_aja-15-0022] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2015] [Accepted: 09/05/2015] [Indexed: 11/09/2022] Open
Abstract
PURPOSE The goal of this work was to better understand speech perception for cochlear implant (CI) users with bilateral residual hearing, including consideration of effects related to listening conditions and test measures. Of interest was the role of acoustic hearing for speech perception in a complex background, the role of listening experience for CI-alone conditions, and whether performance with electric-acoustic stimulation (EAS) was improved by a contralateral hearing aid (HA). METHOD Eleven subjects provided data on Consonant-Nucleus-Consonant (CNC; Peterson & Lehiste, 1962) words in quiet, City University of New York (CUNY; Boothroyd, Hanin, & Hnath, 1985) sentences in steady noise, and Bamford-Kowal-Bench (Bench, Kowal, & Bamford, 1979) sentences in multitalker babble. Listening conditions included: CI with a full-frequency map, CI with a truncated-frequency map, EAS, and EAS+HA (EAS plus contralateral HA). Sounds were presented at 0° azimuth. RESULTS For CNC words and CUNY sentences, performance was better with the truncated-frequency than the full-frequency map, and performance with EAS was better than for either CI-alone condition. For Bench-Kowal-Bamford sentences, EAS+HA was better than EAS. CONCLUSIONS As demonstrated previously, performance was better in the EAS condition than either CI-alone condition. Better performance in the truncated-frequency than full-frequency CI-alone condition suggests that listening experience may be important. A contralateral HA improved performance over unilateral EAS under some conditions.
Collapse
Affiliation(s)
| | - Emily Buss
- University of North Carolina at Chapel Hill
| | | | | | | |
Collapse
|
41
|
Aguiar DE, Taylor NE, Li J, Gazanfari DK, Talavage TM, Laflen JB, Neuberger H, Svirsky MA. Information theoretic evaluation of a noiseband-based cochlear implant simulator. Hear Res 2015; 333:185-193. [PMID: 26409068 DOI: 10.1016/j.heares.2015.09.008] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/02/2015] [Revised: 08/25/2015] [Accepted: 09/20/2015] [Indexed: 10/23/2022]
Abstract
Noise-band vocoders are often used to simulate the signal processing algorithms used in cochlear implants (CIs), producing acoustic stimuli that may be presented to normal hearing (NH) subjects. Such evaluations may obviate the heterogeneity of CI user populations, achieving greater experimental control than when testing on CI subjects. However, it remains an open question whether advancements in algorithms developed on NH subjects using a simulator will necessarily improve performance in CI users. This study assessed the similarity in vowel identification of CI subjects and NH subjects using an 8-channel noise-band vocoder simulator configured to match input and output frequencies or to mimic output after a basalward shift of input frequencies. Under each stimulus condition, NH subjects performed the task both with and without feedback/training. Similarity of NH subjects to CI users was evaluated using correct identification rates and information theoretic approaches. Feedback/training produced higher rates of correct identification, as expected, but also resulted in error patterns that were closer to those of the CI users. Further evaluation remains necessary to determine how patterns of confusion at the token level are affected by the various parameters in CI simulators, providing insight into how a true CI simulation may be developed to facilitate more rapid prototyping and testing of novel CI signal processing and electrical stimulation strategies.
Collapse
Affiliation(s)
- Daniel E Aguiar
- School of Electrical & Computer Engineering, Purdue University, West Lafayette, IN, USA
| | - N Ellen Taylor
- Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN, USA
| | - Jing Li
- School of Electrical & Computer Engineering, Purdue University, West Lafayette, IN, USA
| | - Daniel K Gazanfari
- School of Electrical & Computer Engineering, Purdue University, West Lafayette, IN, USA
| | - Thomas M Talavage
- School of Electrical & Computer Engineering, Purdue University, West Lafayette, IN, USA; Weldon School of Biomedical Engineering, Purdue University, West Lafayette, IN, USA.
| | - J Brandon Laflen
- School of Electrical & Computer Engineering, Purdue University, West Lafayette, IN, USA
| | - Heidi Neuberger
- DeVault Otologic Research Laboratory, Department of Otolaryngology/Head and Neck Surgery, Indiana University School of Medicine, Indianapolis, IN, USA
| | - Mario A Svirsky
- DeVault Otologic Research Laboratory, Department of Otolaryngology/Head and Neck Surgery, Indiana University School of Medicine, Indianapolis, IN, USA; Department of Otolaryngology-Head & Neck Surgery, New York University School of Medicine, New York, NY, USA
| |
Collapse
|
42
|
Azadpour M, Balaban E. A proposed mechanism for rapid adaptation to spectrally distorted speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 138:44-57. [PMID: 26233005 DOI: 10.1121/1.4922226] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
The mechanisms underlying perceptual adaptation to severely spectrally-distorted speech were studied by training participants to comprehend spectrally-rotated speech, which is obtained by inverting the speech spectrum. Spectral-rotation produces severe distortion confined to the spectral domain while preserving temporal trajectories. During five 1-hour training sessions, pairs of participants attempted to extract spoken messages from the spectrally-rotated speech of their training partner. Data on training-induced changes in comprehension of spectrally-rotated sentences and identification/discrimination of spectrally-rotated phonemes were used to evaluate the plausibility of three different classes of underlying perceptual mechanisms: (1) phonemic remapping (the formation of new phonemic categories that specifically incorporate spectrally-rotated acoustic information); (2) experience-dependent generation of a perceptual "inverse-transform" that compensates for spectral-rotation; and (3) changes in cue weighting (the identification of sets of acoustic cues least affected by spectral-rotation, followed by a rapid shift in perceptual emphasis to favour those cues, combined with the recruitment of the same type of "perceptual filling-in" mechanisms used to disambiguate speech-in-noise). Results exclusively support the third mechanism, which is the only one predicting that learning would specifically target temporally-dynamic cues that were transmitting phonetic information most stably in spite of spectral-distortion. No support was found for phonemic remapping or for inverse-transform generation.
Collapse
Affiliation(s)
- Mahan Azadpour
- Cognitive Neuroscience Sector, SISSA (International School for Advanced Studies), Via Beirut 2-4, Trieste, Italy
| | - Evan Balaban
- Cognitive Neuroscience Sector, SISSA (International School for Advanced Studies), Via Beirut 2-4, Trieste, Italy
| |
Collapse
|
43
|
Scheperle RA, Abbas PJ. Relationships Among Peripheral and Central Electrophysiological Measures of Spatial and Spectral Selectivity and Speech Perception in Cochlear Implant Users. Ear Hear 2015; 36:441-53. [PMID: 25658746 PMCID: PMC4478147 DOI: 10.1097/aud.0000000000000144] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES The ability to perceive speech is related to the listener's ability to differentiate among frequencies (i.e., spectral resolution). Cochlear implant (CI) users exhibit variable speech-perception and spectral-resolution abilities, which can be attributed in part to the extent of electrode interactions at the periphery (i.e., spatial selectivity). However, electrophysiological measures of peripheral spatial selectivity have not been found to correlate with speech perception. The purpose of this study was to evaluate auditory processing at the periphery and cortex using both simple and spectrally complex stimuli to better understand the stages of neural processing underlying speech perception. The hypotheses were that (1) by more completely characterizing peripheral excitation patterns than in previous studies, significant correlations with measures of spectral selectivity and speech perception would be observed, (2) adding information about processing at a level central to the auditory nerve would account for additional variability in speech perception, and (3) responses elicited with spectrally complex stimuli would be more strongly correlated with speech perception than responses elicited with spectrally simple stimuli. DESIGN Eleven adult CI users participated. Three experimental processor programs (MAPs) were created to vary the likelihood of electrode interactions within each participant. For each MAP, a subset of 7 of 22 intracochlear electrodes was activated: adjacent (MAP 1), every other (MAP 2), or every third (MAP 3). Peripheral spatial selectivity was assessed using the electrically evoked compound action potential (ECAP) to obtain channel-interaction functions for all activated electrodes (13 functions total). Central processing was assessed by eliciting the auditory change complex with both spatial (electrode pairs) and spectral (rippled noise) stimulus changes. Speech-perception measures included vowel discrimination and the Bamford-Kowal-Bench Speech-in-Noise test. Spatial and spectral selectivity and speech perception were expected to be poorest with MAP 1 (closest electrode spacing) and best with MAP 3 (widest electrode spacing). Relationships among the electrophysiological and speech-perception measures were evaluated using mixed-model and simple linear regression analyses. RESULTS All electrophysiological measures were significantly correlated with each other and with speech scores for the mixed-model analysis, which takes into account multiple measures per person (i.e., experimental MAPs). The ECAP measures were the best predictor. In the simple linear regression analysis on MAP 3 data, only the cortical measures were significantly correlated with speech scores; spectral auditory change complex amplitude was the strongest predictor. CONCLUSIONS The results suggest that both peripheral and central electrophysiological measures of spatial and spectral selectivity provide valuable information about speech perception. Clinically, it is often desirable to optimize performance for individual CI users. These results suggest that ECAP measures may be most useful for within-subject applications when multiple measures are performed to make decisions about processor options. They also suggest that if the goal is to compare performance across individuals based on a single measure, then processing central to the auditory nerve (specifically, cortical measures of discriminability) should be considered.
Collapse
Affiliation(s)
- Rachel A. Scheperle
- Department of Communication Sciences and Disorders, University of Iowa, Iowa
City, IA, USA
| | - Paul J. Abbas
- Department of Communication Sciences and Disorders, University of Iowa, Iowa
City, IA, USA
- Otolaryngology-Head and Neck Surgery, University of Iowa, Iowa City, IA,
USA
| |
Collapse
|
44
|
Chatterjee M, Zion DJ, Deroche ML, Burianek BA, Limb CJ, Goren AP, Kulkarni AM, Christensen JA. Voice emotion recognition by cochlear-implanted children and their normally-hearing peers. Hear Res 2015; 322:151-62. [PMID: 25448167 PMCID: PMC4615700 DOI: 10.1016/j.heares.2014.10.003] [Citation(s) in RCA: 97] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/17/2014] [Revised: 08/27/2014] [Accepted: 10/06/2014] [Indexed: 10/24/2022]
Abstract
Despite their remarkable success in bringing spoken language to hearing impaired listeners, the signal transmitted through cochlear implants (CIs) remains impoverished in spectro-temporal fine structure. As a consequence, pitch-dominant information such as voice emotion, is diminished. For young children, the ability to correctly identify the mood/intent of the speaker (which may not always be visible in their facial expression) is an important aspect of social and linguistic development. Previous work in the field has shown that children with cochlear implants (cCI) have significant deficits in voice emotion recognition relative to their normally hearing peers (cNH). Here, we report on voice emotion recognition by a cohort of 36 school-aged cCI. Additionally, we provide for the first time, a comparison of their performance to that of cNH and NH adults (aNH) listening to CI simulations of the same stimuli. We also provide comparisons to the performance of adult listeners with CIs (aCI), most of whom learned language primarily through normal acoustic hearing. Results indicate that, despite strong variability, on average, cCI perform similarly to their adult counterparts; that both groups' mean performance is similar to aNHs' performance with 8-channel noise-vocoded speech; that cNH achieve excellent scores in voice emotion recognition with full-spectrum speech, but on average, show significantly poorer scores than aNH with 8-channel noise-vocoded speech. A strong developmental effect was observed in the cNH with noise-vocoded speech in this task. These results point to the considerable benefit obtained by cochlear-implanted children from their devices, but also underscore the need for further research and development in this important and neglected area. This article is part of a Special Issue entitled .
Collapse
Affiliation(s)
- Monita Chatterjee
- Auditory Prostheses & Perception Lab., Boys Town National Research Hospital, 555 N 30th St, Omaha, NE 68131, USA.
| | - Danielle J Zion
- Department of Hearing & Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, MD 20742, USA
| | - Mickael L Deroche
- Department of Otolaryngology, Johns Hopkins University School of Medicine, 818 Ross Research Building, 720 Rutland Avenue, Baltimore, MD, USA
| | - Brooke A Burianek
- Auditory Prostheses & Perception Lab., Boys Town National Research Hospital, 555 N 30th St, Omaha, NE 68131, USA
| | - Charles J Limb
- Department of Otolaryngology, Johns Hopkins University School of Medicine, 818 Ross Research Building, 720 Rutland Avenue, Baltimore, MD, USA
| | - Alison P Goren
- Auditory Prostheses & Perception Lab., Boys Town National Research Hospital, 555 N 30th St, Omaha, NE 68131, USA; Department of Hearing & Speech Sciences, University of Maryland, 0100 LeFrak Hall, College Park, MD 20742, USA
| | - Aditya M Kulkarni
- Auditory Prostheses & Perception Lab., Boys Town National Research Hospital, 555 N 30th St, Omaha, NE 68131, USA
| | - Julie A Christensen
- Auditory Prostheses & Perception Lab., Boys Town National Research Hospital, 555 N 30th St, Omaha, NE 68131, USA
| |
Collapse
|
45
|
Shannon RV. Auditory implant research at the House Ear Institute 1989-2013. Hear Res 2015; 322:57-66. [PMID: 25449009 PMCID: PMC4380593 DOI: 10.1016/j.heares.2014.11.003] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/11/2014] [Revised: 11/04/2014] [Accepted: 11/07/2014] [Indexed: 11/29/2022]
Abstract
The House Ear Institute (HEI) had a long and distinguished history of auditory implant innovation and development. Early clinical innovations include being one of the first cochlear implant (CI) centers, being the first center to implant a child with a cochlear implant in the US, developing the auditory brainstem implant, and developing multiple surgical approaches and tools for Otology. This paper reviews the second stage of auditory implant research at House - in-depth basic research on perceptual capabilities and signal processing for both cochlear implants and auditory brainstem implants. Psychophysical studies characterized the loudness and temporal perceptual properties of electrical stimulation as a function of electrical parameters. Speech studies with the noise-band vocoder showed that only four bands of tonotopically arrayed information were sufficient for speech recognition, and that most implant users were receiving the equivalent of 8-10 bands of information. The noise-band vocoder allowed us to evaluate the effects of the manipulation of the number of bands, the alignment of the bands with the original tonotopic map, and distortions in the tonotopic mapping, including holes in the neural representation. Stimulation pulse rate was shown to have only a small effect on speech recognition. Electric fields were manipulated in position and sharpness, showing the potential benefit of improved tonotopic selectivity. Auditory training shows great promise for improving speech recognition for all patients. And the Auditory Brainstem Implant was developed and improved and its application expanded to new populations. Overall, the last 25 years of research at HEI helped increase the basic scientific understanding of electrical stimulation of hearing and contributed to the improved outcomes for patients with the CI and ABI devices. This article is part of a Special Issue entitled .
Collapse
Affiliation(s)
- Robert V Shannon
- Department of Otolaryngology, University of Southern California, Keck School of Medicine of USC, 806 W. Adams Blvd, Los Angeles, CA 90007-2505, USA.
| |
Collapse
|
46
|
Won JH, Jones GL, Moon IJ, Rubinstein JT. Spectral and temporal analysis of simulated dead regions in cochlear implants. J Assoc Res Otolaryngol 2015; 16:285-307. [PMID: 25740402 DOI: 10.1007/s10162-014-0502-8] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2014] [Accepted: 12/23/2014] [Indexed: 11/29/2022] Open
Abstract
A cochlear implant (CI) electrode in a "cochlear dead region" will excite neighboring neural populations. In previous research that simulated such dead regions, stimulus information in the simulated dead region was either added to the immediately adjacent frequency regions or dropped entirely. There was little difference in speech perception ability between the two conditions. This may imply that there may be little benefit of ensuring that stimulus information on an electrode in a suspected cochlear dead region is transmitted. Alternatively, performance may be enhanced by a broader frequency redistribution, rather than adding stimuli from the dead region to the edges. In the current experiments, cochlear dead regions were introduced by excluding selected CI electrodes or vocoder noise-bands. Participants were assessed for speech understanding as well as spectral and temporal sensitivities as a function of the size of simulated dead regions. In one set of tests, the normal input frequency range of the sound processor was distributed among the active electrodes in bands with approximately logarithmic spacing ("redistributed" maps); in the remaining tests, information in simulated dead regions was dropped ("dropped" maps). Word recognition and Schroeder-phase discrimination performance, which require both spectral and temporal sensitivities, decreased as the size of simulated dead regions increased, but the redistributed and dropped remappings showed similar performance in these two tasks. Psychoacoustic experiments showed that the near match in word scores may reflect a tradeoff between spectral and temporal sensitivity: spectral-ripple discrimination was substantially degraded in the redistributed condition relative to the dropped condition while performance in a temporal modulation detection task degraded in the dropped condition but remained constant in the redistributed condition.
Collapse
Affiliation(s)
- Jong Ho Won
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, WA, 98195, USA
| | | | | | | |
Collapse
|
47
|
Grasmeder ML, Verschuur CA, Batty VB. Optimizing frequency-to-electrode allocation for individual cochlear implant users. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:3313. [PMID: 25480076 DOI: 10.1121/1.4900831] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/20/2023]
Abstract
Individual adjustment of frequency-to-electrode assignment in cochlear implants (CIs) may potentially improve speech perception outcomes. Twelve adult CI users were recruited for an experiment, in which frequency maps were adjusted using insertion angles estimated from post-operative x rays; results were analyzed for ten participants with good quality x rays. The allocations were a mapping to the Greenwood function, a compressed map limited to the area containing spiral ganglion (SG) cells, a reduced frequency range map (RFR), and participants' clinical maps. A trial period of at least six weeks was given for the clinical, Greenwood, and SG maps although participants could return to their clinical map if they wished. Performance with the Greenwood map was poor for both sentence and vowel perception and correlated with insertion angle; performance with the SG map was poorer than for the clinical map. The RFR map was significantly better than the clinical map for three participants, for sentence perception, but worse for three others. Those with improved performance had relatively deep insertions and poor electrode discrimination ability for apical electrodes. The results suggest that CI performance could be improved by adjustment of the frequency allocation, based on a measure of insertion angle and/or electrode discrimination ability.
Collapse
Affiliation(s)
- Mary L Grasmeder
- Auditory Implant Service, Faculty of Engineering and the Environment, Building 19, University of Southampton, Southampton SO17 1BJ, United Kingdom
| | - Carl A Verschuur
- Auditory Implant Service, Faculty of Engineering and the Environment, Building 19, University of Southampton, Southampton SO17 1BJ, United Kingdom
| | - Vincent B Batty
- Wessex Neurological Centre, University Hospital Southampton NHS Foundation Trust, Tremona Road, Southampton SO16 6YD, United Kingdom
| |
Collapse
|
48
|
Mesnildrey Q, Macherey O. Simulating the dual-peak excitation pattern produced by bipolar stimulation of a cochlear implant: effects on speech intelligibility. Hear Res 2014; 319:32-47. [PMID: 25449010 DOI: 10.1016/j.heares.2014.11.001] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/19/2014] [Revised: 10/28/2014] [Accepted: 11/05/2014] [Indexed: 10/24/2022]
Abstract
Several electrophysiological and psychophysical studies have shown that the spatial excitation pattern produced by bipolar stimulation of a cochlear implant (CI) can have a dual-peak shape. The perceptual effects of this dual-peak shape were investigated using noise-vocoded CI simulations in which synthesis filters were designed to simulate the spread of neural activity produced by various electrode configurations, as predicted by a simple cochlear model. Experiments 1 and 2 tested speech recognition in the presence of a concurrent speech masker for various sets of single-peak and dual-peak synthesis filters and different numbers of channels. Similarly as results obtained in real CIs, both monopolar (MP, single-peak) and bipolar (BP + 1, dual-peak) simulations showed a plateau of performance above 8 channels. The benefit of increasing the number of channels was also lower for BP + 1 than for MP. This shows that channel interactions in BP + 1 become especially deleterious for speech intelligibility when a simulated electrode acts both as an active and as a return electrode for different channels because envelope information from two different analysis bands are being conveyed to the same spectral location. Experiment 3 shows that these channel interactions are even stronger in wide BP configuration (BP + 5), likely because the interfering speech envelopes are less correlated than in narrow BP + 1. Although the exact effects of dual- or multi-peak excitation in real CIs remain to be determined, this series of experiments suggest that multipolar stimulation strategies, such as bipolar or tripolar, should be controlled to avoid neural excitation in the vicinity of the return electrodes.
Collapse
Affiliation(s)
- Quentin Mesnildrey
- LMA-CNRS, UPR 7051, Aix-Marseille Univ., Centrale Marseille, 31 Chemin Joseph Aiguier, 13402 Marseille Cedex 20, France.
| | - Olivier Macherey
- LMA-CNRS, UPR 7051, Aix-Marseille Univ., Centrale Marseille, 31 Chemin Joseph Aiguier, 13402 Marseille Cedex 20, France
| |
Collapse
|
49
|
Svirsky MA, Talavage TM, Sinha S, Neuburger H, Azadpour M. Gradual adaptation to auditory frequency mismatch. Hear Res 2014; 322:163-70. [PMID: 25445816 DOI: 10.1016/j.heares.2014.10.008] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/13/2014] [Revised: 10/13/2014] [Accepted: 10/16/2014] [Indexed: 12/01/2022]
Abstract
What is the best way to help humans adapt to a distorted sensory input? Interest in this question is more than academic. The answer may help facilitate auditory learning by people who became deaf after learning language and later received a cochlear implant (a neural prosthesis that restores hearing through direct electrical stimulation of the auditory nerve). There is evidence that some cochlear implants (which provide information that is spectrally degraded to begin with) stimulate neurons with higher characteristic frequency than the acoustic frequency of the original stimulus. In other words, the stimulus is shifted in frequency with respect to what the listener expects to hear. This frequency misalignment may have a negative influence on speech perception by CI users. However, a perfect frequency-place alignment may result in the loss of important low frequency speech information. A trade-off may involve a gradual approach: start with correct frequency-place alignment to allow listeners to adapt to the spectrally degraded signal first, and then gradually increase the frequency shift to allow them to adapt to it over time. We used an acoustic model of a cochlear implant to measure adaptation to a frequency-shifted signal, using either the gradual approach or the "standard" approach (sudden imposition of the frequency shift). Listeners in both groups showed substantial auditory learning, as measured by increases in speech perception scores over the course of fifteen one-hour training sessions. However, the learning process was faster for listeners who were exposed to the gradual approach. These results suggest that gradual rather than sudden exposure may facilitate perceptual learning in the face of a spectrally degraded, frequency-shifted input. This article is part of a Special Issue entitled <Lasker Award>.
Collapse
Affiliation(s)
- Mario A Svirsky
- Dept. of Otolaryngology-HNS, New York University School of Medicine, New York, NY, USA; Center of Neural Science, New York University, New York, NY, USA.
| | - Thomas M Talavage
- ECE, Purdue University, West Lafayette, IN, USA; BME Depts., Purdue University, West Lafayette, IN, USA
| | | | - Heidi Neuburger
- Dept. of Otolaryngology-HNS, Indiana University School of Medicine, Indianapolis, IN, USA
| | - Mahan Azadpour
- Dept. of Otolaryngology-HNS, New York University School of Medicine, New York, NY, USA
| |
Collapse
|
50
|
Plant KL, McDermott HJ, van Hoesel RJM, Dawson PW, Cowan RS. Factors influencing electrical place pitch perception in bimodal listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:1199. [PMID: 25190394 DOI: 10.1121/1.4892790] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Factors that might affect perceptual pitch match between acoustic and electric stimulation were examined in 25 bimodal listeners using magnitude estimation. Pre-operative acoustic thresholds in both ears, and duration of severe-profound loss, were first examined as correlates with degree of match between the measured pitch and that predicted by the spiral ganglion frequency-position model. The degree of match was examined with respect to (1) the ratio between the measured and predicted pitch percept on the most apical electrode and (2) the ratio between the slope of the measured and predicted pitch function. Second, effect of listening experience was examined to assess whether adaptation occurred over time to match the frequency assignment to electrodes. Pre-experience pitch estimates on the apical electrode were within the predicted range in only 28% of subjects, and the slope of the electrical pitch function was lower than predicted in all except one subject. Subjects with poorer hearing tended to have a lower pitch and a shallower electrical pitch function than predicted by the model. Pre-operative hearing thresholds in the contralateral ear and hearing loss duration were not correlated with the degree of pitch match, and there was no significant group effect of listening experience.
Collapse
Affiliation(s)
- Kerrie L Plant
- The HEARing CRC, Level 1, 174 Victoria Parade, East Melbourne, Victoria 3002, Australia
| | - Hugh J McDermott
- Bionics Institute, 384-388 Albert Street, East Melbourne, Victoria 3002, Australia
| | | | - Pamela W Dawson
- The HEARing CRC, 550 Swanston Street, Parkville, Victoria 3010, Australia
| | - Robert S Cowan
- The HEARing CRC, 550 Swanston Street, Parkville, Victoria 3010, Australia
| |
Collapse
|