1
|
Drouin JR, Flores S. Effects of training length on adaptation to noise-vocoded speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2114-2127. [PMID: 38488452 DOI: 10.1121/10.0025273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2023] [Accepted: 02/22/2024] [Indexed: 03/19/2024]
Abstract
Listeners show rapid perceptual learning of acoustically degraded speech, though the amount of exposure required to maximize speech adaptation is unspecified. The current work used a single-session design to examine the length of auditory training on perceptual learning for normal hearing listeners exposed to eight-channel noise-vocoded speech. Participants completed short, medium, or long training using a two-alternative forced choice sentence identification task with feedback. To assess learning and generalization, a 40-trial pre-test and post-test transcription task was administered using trained and novel sentences. Training results showed all groups performed near ceiling with no reliable differences. For test data, we evaluated changes in transcription accuracy using separate linear mixed models for trained or novel sentences. In both models, we observed a significant improvement in transcription at post-test relative to pre-test. Critically, the three training groups did not differ in the magnitude of improvement following training. Subsequent Bayes factors analysis evaluating the test by group interaction provided strong evidence in support of the null hypothesis. For these stimuli and procedure, results suggest increased training does not necessarily maximize learning outcomes; both passive and trained experience likely supported adaptation. Findings may contribute to rehabilitation recommendations for listeners adapting to degraded speech signals.
Collapse
Affiliation(s)
- Julia R Drouin
- Division of Speech and Hearing Sciences, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA
| | - Stephany Flores
- Department of Communication Sciences and Disorders, California State University Fullerton, Fullerton, California 92831, USA
| |
Collapse
|
2
|
Pieper SH, Hamze N, Brill S, Hochmuth S, Exter M, Polak M, Radeloff A, Buschermöhle M, Dietz M. Considerations for Fitting Cochlear Implants Bimodally and to the Single-Sided Deaf. Trends Hear 2022; 26:23312165221108259. [PMID: 35726211 PMCID: PMC9218456 DOI: 10.1177/23312165221108259] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
When listening with a cochlear implant through one ear and acoustically through the other, binaural benefits and spatial hearing abilities are generally poorer than in other bilaterally stimulated configurations. With the working hypothesis that binaural neurons require interaurally matched inputs, we review causes for mismatch, their perceptual consequences, and experimental methods for mismatch measurements. The focus is on the three primary interaural dimensions of latency, frequency, and level. Often, the mismatch is not constant, but rather highly stimulus-dependent. We report on mismatch compensation strategies, taking into consideration the specific needs of the respective patient groups. Practical challenges typically faced by audiologists in the proposed fitting procedure are discussed. While improvement in certain areas (e.g., speaker localization) is definitely achievable, a more comprehensive mismatch compensation is a very ambitious endeavor. Even in the hypothetical ideal fitting case, performance is not expected to exceed that of a good bilateral cochlear implant user.
Collapse
Affiliation(s)
- Sabrina H Pieper
- Department of Medical Physics and Acoustic, University of Oldenburg, Oldenburg, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany
| | - Noura Hamze
- MED-EL Medical Electronics GmbH, Innsbruck, Austria
| | - Stefan Brill
- MED-EL Medical Electronics Germany GmbH, Starnberg, Germany
| | - Sabine Hochmuth
- Division of Otorhinolaryngology, University of Oldenburg, Oldenburg, Germany
| | - Mats Exter
- Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Hörzentrum Oldenburg gGmbH, Oldenburg, Germany
| | - Marek Polak
- MED-EL Medical Electronics GmbH, Innsbruck, Austria
| | - Andreas Radeloff
- Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Division of Otorhinolaryngology, University of Oldenburg, Oldenburg, Germany.,Research Center Neurosensory Science, University of Oldenburg, Oldenburg, Germany
| | | | - Mathias Dietz
- Department of Medical Physics and Acoustic, University of Oldenburg, Oldenburg, Germany.,Cluster of Excellence Hearing4all, University of Oldenburg, Oldenburg, Germany.,Research Center Neurosensory Science, University of Oldenburg, Oldenburg, Germany
| |
Collapse
|
3
|
Wright NC, Borden PY, Liew YJ, Bolus MF, Stoy WM, Forest CR, Stanley GB. Rapid Cortical Adaptation and the Role of Thalamic Synchrony during Wakefulness. J Neurosci 2021; 41:5421-5439. [PMID: 33986072 PMCID: PMC8221593 DOI: 10.1523/jneurosci.3018-20.2021] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Revised: 03/18/2021] [Accepted: 04/29/2021] [Indexed: 12/14/2022] Open
Abstract
Rapid sensory adaptation is observed across all sensory systems, and strongly shapes sensory percepts in complex sensory environments. Yet despite its ubiquity and likely necessity for survival, the mechanistic basis is poorly understood. A wide range of primarily in vitro and anesthetized studies have demonstrated the emergence of adaptation at the level of primary sensory cortex, with only modest signatures in earlier stages of processing. The nature of rapid adaptation and how it shapes sensory representations during wakefulness, and thus the potential role in perceptual adaptation, is underexplored, as are the mechanisms that underlie this phenomenon. To address these knowledge gaps, we recorded spiking activity in primary somatosensory cortex (S1) and the upstream ventral posteromedial (VPm) thalamic nucleus in the vibrissa pathway of awake male and female mice, and quantified responses to whisker stimuli delivered in isolation and embedded in an adapting sensory background. We found that cortical sensory responses were indeed adapted by persistent sensory stimulation; putative excitatory neurons were profoundly adapted, and inhibitory neurons only modestly so. Further optogenetic manipulation experiments and network modeling suggest this largely reflects adaptive changes in synchronous thalamic firing combined with robust engagement of feedforward inhibition, with little contribution from synaptic depression. Taken together, these results suggest that cortical adaptation in the regime explored here results from changes in the timing of thalamic input, and the way in which this differentially impacts cortical excitation and feedforward inhibition, pointing to a prominent role of thalamic gating in rapid adaptation of primary sensory cortex.SIGNIFICANCE STATEMENT Rapid adaptation of sensory activity strongly shapes representations of sensory inputs across all sensory pathways over the timescale of seconds, and has profound effects on sensory perception. Despite its ubiquity and theoretical role in the efficient encoding of complex sensory environments, the mechanistic basis is poorly understood, particularly during wakefulness. In this study in the vibrissa pathway of awake mice, we show that cortical representations of sensory inputs are strongly shaped by rapid adaptation, and that this is mediated primarily by adaptive gating of the thalamic inputs to primary sensory cortex and the differential way in which these inputs engage cortical subpopulations of neurons.
Collapse
Affiliation(s)
- Nathaniel C Wright
- Wallace H Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, Georgia 30332
| | - Peter Y Borden
- Wallace H Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, Georgia 30332
| | - Yi Juin Liew
- Department of Biomedical Engineering, Georgia Institute of Technology, Emory University, Atlanta, Georgia 30332 and Beijing University, Beijing China 100871
| | - Michael F Bolus
- Wallace H Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, Georgia 30332
| | - William M Stoy
- Department of Mechanical Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332
| | - Craig R Forest
- Department of Mechanical Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332
| | - Garrett B Stanley
- Wallace H Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, Georgia 30332
| |
Collapse
|
4
|
Abstract
The most important parameter that affects the ability to hear and understand speech in the presence of background noise is the signal-to-noise ratio (SNR). Despite decades of research in speech intelligibility, it is not currently known how much improvement in SNR is needed to provide a meaningful benefit to someone. We propose that the underlying psychophysical basis to a meaningful benefit should be the just noticeable difference (JND) for SNR. The SNR JND was measured in a series of experiments using both adaptive and fixed-level procedures across participants of varying hearing ability. The results showed an average SNR JND of approximately 3 dB for sentences in same-spectrum noise. The role of the stimulus and link to intelligibility was examined by measuring speech-intelligibility psychometric functions and comparing the intelligibility JND estimated from those functions with measured SNR JNDs. Several experiments were then conducted to establish a just meaningful difference (JMD) for SNR. SNR changes that could induce intervention-seeking behaviour for an individual were measured with subjective scaling and report, using the same stimuli as the SNR JND experiment as pre- and post-benefit examples. The results across different rating and willingness-to-change tasks showed that the mean ratings increased near linearly with a change in SNR, but a change of at least 6 dB was necessary to reliably motivate participants to seek intervention. The magnitude of the JNDs and JMDs for speech-intelligibility benefits measured here suggest a gap between what is achievable and what is meaningful.
Collapse
|
5
|
McShefferty D, Whitmer WM, Akeroyd MA. The Just-Meaningful Difference in Speech-to-Noise Ratio. Trends Hear 2016; 20:20/0/2331216515626570. [PMID: 26834121 PMCID: PMC4737977 DOI: 10.1177/2331216515626570] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The speech-to-noise ratio (SNR) in an environment plays a vital role in speech communication for both normal-hearing (NH) and hearing-impaired (HI) listeners. While hearing-assistance devices attempt to deliver as favorable an SNR as possible, there may be discrepancies between noticeable and meaningful improvements in SNR. Furthermore, it is not clear how much of an SNR improvement is necessary to induce intervention-seeking behavior. Here, we report on a series of experiments examining the just-meaningful difference (JMD) in SNR. All experiments used sentences in same-spectrum noise, with two intervals on each trial mimicking examples of pre- and post-benefit situations. Different groups of NH and HI adults were asked (a) to rate how much better or worse the change in SNR was in a number of paired examples, (b) if they would swap the worse for the better SNR (e.g., their current device for another), or (c) if they would be willing to go to the clinic for the given increase in SNR. The mean SNR JMD based on better or worse ratings (one arbitrary unit) was similar to the just-noticeable difference, approximately 3 dB. However, the mean SNR JMD for the more clinically relevant tasks-willingness (at least 50% of the time) to swap devices or attend the clinic for a change in SNR--was 6 to 8 dB regardless of hearing ability. This SNR JMD of the order of 6 dB provides a new benchmark, indicating the SNR improvement necessary to immediately motivate participants to seek intervention.
Collapse
Affiliation(s)
- David McShefferty
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, UK
| | - William M Whitmer
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, UK
| | | |
Collapse
|
6
|
Won JH, Jones GL, Moon IJ, Rubinstein JT. Spectral and temporal analysis of simulated dead regions in cochlear implants. J Assoc Res Otolaryngol 2015; 16:285-307. [PMID: 25740402 DOI: 10.1007/s10162-014-0502-8] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2014] [Accepted: 12/23/2014] [Indexed: 11/29/2022] Open
Abstract
A cochlear implant (CI) electrode in a "cochlear dead region" will excite neighboring neural populations. In previous research that simulated such dead regions, stimulus information in the simulated dead region was either added to the immediately adjacent frequency regions or dropped entirely. There was little difference in speech perception ability between the two conditions. This may imply that there may be little benefit of ensuring that stimulus information on an electrode in a suspected cochlear dead region is transmitted. Alternatively, performance may be enhanced by a broader frequency redistribution, rather than adding stimuli from the dead region to the edges. In the current experiments, cochlear dead regions were introduced by excluding selected CI electrodes or vocoder noise-bands. Participants were assessed for speech understanding as well as spectral and temporal sensitivities as a function of the size of simulated dead regions. In one set of tests, the normal input frequency range of the sound processor was distributed among the active electrodes in bands with approximately logarithmic spacing ("redistributed" maps); in the remaining tests, information in simulated dead regions was dropped ("dropped" maps). Word recognition and Schroeder-phase discrimination performance, which require both spectral and temporal sensitivities, decreased as the size of simulated dead regions increased, but the redistributed and dropped remappings showed similar performance in these two tasks. Psychoacoustic experiments showed that the near match in word scores may reflect a tradeoff between spectral and temporal sensitivity: spectral-ripple discrimination was substantially degraded in the redistributed condition relative to the dropped condition while performance in a temporal modulation detection task degraded in the dropped condition but remained constant in the redistributed condition.
Collapse
Affiliation(s)
- Jong Ho Won
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, WA, 98195, USA
| | | | | | | |
Collapse
|
7
|
McShefferty D, Whitmer WM, Akeroyd MA. The just-noticeable difference in speech-to-noise ratio. Trends Hear 2015; 19:19/0/2331216515572316. [PMID: 25681327 PMCID: PMC4335553 DOI: 10.1177/2331216515572316] [Citation(s) in RCA: 26] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Just-noticeable differences (JNDs) have been measured for various features of sounds, but despite its importance to communication, there is no benchmark for what is a just-noticeable-and possibly meaningful-difference in speech-to-noise ratio (SNR). SNR plays a crucial role in speech communication for normal-hearing and hearing-impaired listeners. Difficulty hearing speech in background noise-a poor SNR-often leads to dissatisfaction with hearing-assistance devices. While such devices attempt through various strategies to address this problem, it is not currently known how much improvement in SNR is needed to provide a noticeable benefit. To investigate what is a noticeable benefit, we measured the JND in SNR for both normal-hearing and hearing-impaired listeners. Here, we report the SNR JNDs of 69 participants of varying hearing ability, estimated using either an adaptive or fixed-level procedure. The task was to judge which of the two intervals containing a sentence in speech-spectrum noise presented over headphones was clearer. The level of each interval was roved to reduce the influence of absolute level cues. The results of both procedures showed an average SNR JND of 3 dB that was independent of hearing ability. Further experiments using a subset of normal-hearing listeners showed that level roving does elevate threshold. These results suggest that noise reduction schemes may need to achieve a benefit greater than 3 dB to be reliably discriminable.
Collapse
Affiliation(s)
- David McShefferty
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, Glasgow, UK
| | - William M Whitmer
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, Glasgow, UK
| | - Michael A Akeroyd
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, Glasgow, UK
| |
Collapse
|
8
|
Gresele ADP, Costa MJ, Garcia MV. Compressão de frequências no reconhecimento de fala de idosos com possíveis zonas mortas na cóclea. REVISTA CEFAC 2015. [DOI: 10.1590/1982-021620155414] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
OBJETIVO: avaliar e comparar o desempenho de idosos sem e com zonas mortas na cóclea em testes de reconhecimento de fala, no silêncio e no ruído, usando próteses auditivas sem e com compressão não linear de frequências. MÉTODOS: participaram 38 idosos com perda auditiva de grau leve a moderado e configuração descendente, distribuídos, com base nos resultados da técnica de mascaramento com ruído branco, em: Grupo A - 24 idosos sem indícios de zonas mortas na cóclea; Grupo B - 14 idosos com possíveis zonas mortas na cóclea. Aplicou-se o teste Listas de Sentenças em Português, pesquisando-se os Índices Percentuais de Reconhecimento de Sentenças no Silêncio e no Ruído. As medidas foram obtidas com próteses auditivas, sem e com compressão de frequências. RESULTADOS: o grupo A e o B apresentaram melhora significante no silêncio com as próteses auditivas com compressão de frequências; no ruído nenhum grupo apresentou diferença sem e com compressão de freqüências. Comparando-se os grupos, não houve diferença no silêncio sem e com compressão de frequências. No ruído sem a ativação da compressão houve diferença significante, sendo melhor o desempenho do grupo B. No ruído com a ativação do recurso não houve diferença significante. CONCLUSÃO: no silêncio, ambos os grupos apresentaram melhor desempenho usando próteses com compressão de frequências. No ruído, não houve diferença entre os resultados sem e com compressão de frequências. Comparando-se os grupos, a medida obtida no ruído com próteses auditivas sem compressão de frequências apresentou diferença, na qual o grupo com zonas mortas obteve melhor desempenho.
Collapse
|
9
|
Mendonça C, Campos G, Dias P, Santos JA. Learning auditory space: generalization and long-term effects. PLoS One 2013; 8:e77900. [PMID: 24167588 PMCID: PMC3805533 DOI: 10.1371/journal.pone.0077900] [Citation(s) in RCA: 30] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2013] [Accepted: 09/13/2013] [Indexed: 11/24/2022] Open
Abstract
BACKGROUND Previous findings have shown that humans can learn to localize with altered auditory space cues. Here we analyze such learning processes and their effects up to one month on both localization accuracy and sound externalization. Subjects were trained and retested, focusing on the effects of stimulus type in learning, stimulus type in localization, stimulus position, previous experience, externalization levels, and time. METHOD We trained listeners in azimuth and elevation discrimination in two experiments. Half participated in the azimuth experiment first and half in the elevation first. In each experiment, half were trained in speech sounds and half in white noise. Retests were performed at several time intervals: just after training and one hour, one day, one week and one month later. In a control condition, we tested the effect of systematic retesting over time with post-tests only after training and either one day, one week, or one month later. RESULTS With training all participants lowered their localization errors. This benefit was still present one month after training. Participants were more accurate in the second training phase, revealing an effect of previous experience on a different task. Training with white noise led to better results than training with speech sounds. Moreover, the training benefit generalized to untrained stimulus-position pairs. Throughout the post-tests externalization levels increased. In the control condition the long-term localization improvement was not lower without additional contact with the trained sounds, but externalization levels were lower. CONCLUSION Our findings suggest that humans adapt easily to altered auditory space cues and that such adaptation spreads to untrained positions and sound types. We propose that such learning depends on all available cues, but each cue type might be learned and retrieved differently. The process of localization learning is global, not limited to stimulus-position pairs, and it differs from externalization processes.
Collapse
Affiliation(s)
- Catarina Mendonça
- Cluster of Excellence Hearing4all, Carl Von Ossietzky University of Oldenburg, Oldenburg, Germany
- Department of Signal Processing and Acoustics, Aalto University, Espoo, Finland
- Centro Algoritmi, School of Engineering, University of Minho, Guimarães, Portugal
| | - Guilherme Campos
- Institute of Electronics and Telematics Engineering, University of Aveiro, Aveiro, Portugal
| | - Paulo Dias
- Institute of Electronics and Telematics Engineering, University of Aveiro, Aveiro, Portugal
| | - Jorge A. Santos
- Centro Algoritmi, School of Engineering, University of Minho, Guimarães, Portugal
- Department of Basic Psychology, University of Minho, Braga, Portugal
| |
Collapse
|
10
|
van Besouw RM, Forrester L, Crowe ND, Rowan D. Simulating the effect of interaural mismatch in the insertion depth of bilateral cochlear implants on speech perception. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:1348-1357. [PMID: 23927131 DOI: 10.1121/1.4812272] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
A bilateral advantage for diotically presented stimuli has been observed for cochlear implant (CI) users and is suggested to be dependent on symmetrical implant performance. Studies using CI simulations have not shown a true "bilateral" advantage, but a "better ear" effect and have demonstrated that performance decreases with increasing basalward shift in insertion depth. This study aimed to determine whether there is a bilateral advantage for CI simulations with interaurally matched insertions and the extent to which performance is affected by interaural insertion depth mismatch. Speech perception in noise and self-reported ease of listening were measured using matched bilateral, mismatched bilateral and unilateral CI simulations over four insertion depths for seventeen normal hearing listeners. Speech scores and ease of listening reduced with increasing basalward shift in (interaurally matched) insertion depth. A bilateral advantage for speech perception was only observed when the insertion depths were interaurally matched and deep. No advantage was observed for small to moderate interaural insertion-depth mismatches, consistent with a better ear effect. Finally, both measures were poorer than expected for a better ear effect for large mismatches, suggesting that misalignment of the electrode arrays may prevent a bilateral advantage and detrimentally affect perception of diotically presented speech.
Collapse
Affiliation(s)
- Rachel M van Besouw
- Institute of Sound and Vibration Research, University of Southampton, Southampton, Hampshire, SO17 1BJ, United Kingdom.
| | | | | | | |
Collapse
|
11
|
Green T, Rosen S, Faulkner A, Paterson R. Adaptation to spectrally-rotated speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:1369-1377. [PMID: 23927133 DOI: 10.1121/1.4812759] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
Much recent interest surrounds listeners' abilities to adapt to various transformations that distort speech. An extreme example is spectral rotation, in which the spectrum of low-pass filtered speech is inverted around a center frequency (2 kHz here). Spectral shape and its dynamics are completely altered, rendering speech virtually unintelligible initially. However, intonation, rhythm, and contrasts in periodicity and aperiodicity are largely unaffected. Four normal hearing adults underwent 6 h of training with spectrally-rotated speech using Continuous Discourse Tracking. They and an untrained control group completed pre- and post-training speech perception tests, for which talkers differed from the training talker. Significantly improved recognition of spectrally-rotated sentences was observed for trained, but not untrained, participants. However, there were no significant improvements in the identification of medial vowels in /bVd/ syllables or intervocalic consonants. Additional tests were performed with speech materials manipulated so as to isolate the contribution of various speech features. These showed that preserving intonational contrasts did not contribute to the comprehension of spectrally-rotated speech after training, and suggested that improvements involved adaptation to altered spectral shape and dynamics, rather than just learning to focus on speech features relatively unaffected by the transformation.
Collapse
Affiliation(s)
- Tim Green
- Speech, Hearing, and Phonetic Sciences, UCL, Chandler House, 2, Wakefield Street, London, WC1N 1PF, United Kingdom
| | | | | | | |
Collapse
|
12
|
Jackson HM, Moore BCJ. Contribution of temporal fine structure information and fundamental frequency separation to intelligibility in a competing-speaker paradigm. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:2421-2430. [PMID: 23556607 DOI: 10.1121/1.4792153] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
The speech reception threshold (SRT) for identifying a target speaker in a background speaker was measured as a function of the difference (F0sep) in fundamental frequency (F0) between the two speakers. The amount of original temporal fine structure (TFS) information in the mixed signals was manipulated by tone vocoding channels above a certain cutoff channel (CO). When the natural variations in F0 of both speakers were preserved, the SRT did not decrease with increasing F0sep, indicating that short-term differences in F0 can allow perceptual segregation of two speakers even when their F0s cross. When F0 variations were removed from both speakers, increasing F0sep led to decreased (better) SRTs. The decrease was greater for unprocessed signals than for fully tone-vocoded signals. However, the decrease was similar for unprocessed signals and for signals with original TFS below 1600 Hz, suggesting that most of the benefit from increasing F0 difference depends on the use of TFS information at lower frequencies. Adding original TFS information to channels centered above 1600 Hz produced roughly the same decrease in SRT as adding original TFS information to channels centered below 1600 Hz, suggesting a benefit from original TFS information apart from that related to differences in F0.
Collapse
Affiliation(s)
- Helen M Jackson
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom
| | | |
Collapse
|
13
|
Hilkhuysen GLM, Gaubitch N, Huckvale M. Effects of noise suppression on intelligibility: experts' opinions and naive normal-hearing listeners' performance. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2013; 56:404-415. [PMID: 23090965 DOI: 10.1044/1092-4388(2012/11-0286)] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
PURPOSE In this study, the authors investigated how well experts can adjust the settings of a commercial noise-reduction system to optimize the intelligibility for naive normal-hearing listeners. METHOD In Experiment 1, 5 experts adjusted parameters for a noise-reduction system while aiming to optimize intelligibility. The stimuli consisted of speech presented in car-cabin noise or babble at 5 different signal-to-noise ratios (SNRs). In Experiment 2, the effects of processing with these settings were measured with 10 listeners undertaking an intelligibility test. In Experiment 3, the intelligibility of a broad range of settings was investigated with another 10 listeners to determine whether the experts' chosen settings could have been improved. RESULTS Low Cronbach's alphas indicated that parameter settings varied considerably within and across experts. For very low SNRs, mean proposed settings differed from those for higher SNRs. The different settings had no significant effects on intelligibility for naive normal-hearing listeners. At high SNRs, the settings proposed by experts were found to deteriorate intelligibility. Superior intelligibility for naive normal-hearing listeners was achievable from settings other than the ones proposed by the experts. CONCLUSION While attempting to enhance noisy speech, experts may propose settings that deteriorate intelligibility for naive normal-hearing listeners.
Collapse
|
14
|
Faulkner A, Rosen S, Green T. Comparing live to recorded speech in training the perception of spectrally shifted noise-vocoded speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 132:EL336-EL342. [PMID: 23039574 DOI: 10.1121/1.4754432] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Two experimental groups were trained for 2 h with live or recorded speech that was noise-vocoded and spectrally shifted and was from the same text and talker. These two groups showed equivalent improvements in performance for vocoded and shifted sentences, and the group trained with recorded speech showed consistently greater improvements than untrained controls. Another group trained with unshifted noise-vocoded speech improved no more than untrained controls. Computer-based training thus appears at least as effective as labor-intensive live-voice training for improving the perception of spectrally shifted noise-vocoded speech, and by implication, for training of users of cochlear implants.
Collapse
Affiliation(s)
- Andrew Faulkner
- Speech Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom.
| | | | | |
Collapse
|
15
|
Stone MA, Anton K, Moore BCJ. Use of high-rate envelope speech cues and their perceptually relevant dynamic range for the hearing impaired. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 132:1141-1151. [PMID: 22894233 DOI: 10.1121/1.4733543] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
The ability of hearing-impaired (HI) listeners to use high-rate envelope information in a competing-talker situation was assessed. In experiment 1, signals were tone vocoded and the cutoff frequency (f(c)) of the envelope extraction filter was either 50 Hz (E filter) or 200 Hz (P filter). The channels for which the P or E filter was used were varied. Intelligibility was higher with the P filter regardless of whether it was used for low or high center frequencies. Performance was best when the P filter was used for all channels. Experiment 2 explored the dynamic range over which HI listeners made use of high-rate cues. In each channel of a vocoder, the envelope extracted using f(c) = 16 Hz was replaced by the envelope extracted using f(c) = 300 Hz, either at the peaks or valleys, with a parametrically varied "switching threshold." For a target-to-background ratio of +5 dB, changes in speech intelligibility occurred mainly when the switching threshold was between -8 and +8 dB relative to the channel root-mean-square level. This range is similar in width to, but about 3 dB higher in absolute level than, that found for normal-hearing listeners, despite the reduced dynamic range of the HI listeners.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, England.
| | | | | |
Collapse
|
16
|
Stone MA, Füllgrabe C, Moore BCJ. Notionally steady background noise acts primarily as a modulation masker of speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 132:317-26. [PMID: 22779480 DOI: 10.1121/1.4725766] [Citation(s) in RCA: 55] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
Stone et al. [J. Acoust. Soc Am. 130, 2874-2881 (2011)], using vocoder processing, showed that the envelope modulations of a notionally steady noise were more effective than the envelope energy as a masker of speech. Here the same effect is demonstrated using non-vocoded signals. Speech was filtered into 28 channels. A masker centered on each channel was added to the channel signal at a target-to-background ratio of -5 or -10 dB. Maskers were sinusoids or noise bands with bandwidth 1/3 or 1 ERB(N) (ERB(N) being the bandwidth of "normal" auditory filters), synthesized with Gaussian (GN) or low-noise (LNN) statistics. To minimize peripheral interactions between maskers, odd-numbered channels were presented to one ear and even to the other. Speech intelligibility was assessed in the presence of each "steady" masker and that masker 100% sinusoidally amplitude modulated (SAM) at 8 Hz. Intelligibility decreased with increasing envelope fluctuation of the maskers. Masking release, the difference in intelligibility between the SAM and its "steady" counterpart, increased with bandwidth from near-zero to around 50 percentage points for the 1-ERB(N) GN. It is concluded that the sinusoidal and GN maskers behaved primarily as energetic and modulation maskers, respectively.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom.
| | | | | |
Collapse
|
17
|
Hilkhuysen G, Gaubitch N, Brookes M, Huckvale M. Effects of noise suppression on intelligibility: dependency on signal-to-noise ratios. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2012; 131:531-539. [PMID: 22280614 DOI: 10.1121/1.3665996] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
The effects on speech intelligibility of three different noise reduction algorithms (spectral subtraction, minimal mean squared error spectral estimation, and subspace analysis) were evaluated in two types of noise (car and babble) over a 12 dB range of signal-to-noise ratios (SNRs). Results from these listening experiments showed that most algorithms deteriorated intelligibility scores. Modeling of the results with a logit-shaped psychometric function showed that the degradation in intelligibility scores was largely congruent with a constant shift in SNR, although some additional degradation was observed at two SNRs, suggesting a limited interaction between the effects of noise suppression and SNR.
Collapse
Affiliation(s)
- Gaston Hilkhuysen
- Department of Speech, Hearing and Phonetic Sciences, University College London, 2 Wakefield Street, London WC1N 1PF, United Kingdom.
| | | | | | | |
Collapse
|
18
|
Stone MA, Füllgrabe C, Mackinnon RC, Moore BCJ. The importance for speech intelligibility of random fluctuations in "steady" background noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 130:2874-81. [PMID: 22087916 DOI: 10.1121/1.3641371] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
Spectrally shaped steady noise is commonly used as a masker of speech. The effects of inherent random fluctuations in amplitude of such a noise are typically ignored. Here, the importance of these random fluctuations was assessed by comparing two cases. For one, speech was mixed with steady speech-shaped noise and N-channel tone vocoded, a process referred to as signal-domain mixing (SDM); this preserved the random fluctuations of the noise. For the second, the envelope of speech alone was extracted for each vocoder channel and a constant was added corresponding to the root-mean-square value of the noise envelope for that channel. This is referred to as envelope-domain mixing (EDM); it removed the random fluctuations of the noise. Sinusoidally modulated noise and a single talker were also used as backgrounds, with both SDM and EDM. Speech intelligibility was measured for N = 12, 19, and 30, with the target-to-background ratio fixed at -7 dB. For SDM, performance was best for the speech background and worst for the steady noise. For EDM, this pattern was reversed. Intelligibility with steady noise was consistently very poor for SDM, but near-ceiling for EDM, demonstrating that the random fluctuations in steady noise have a large effect.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, England.
| | | | | | | |
Collapse
|
19
|
Stone MA, Moore BCJ, Füllgrabe C. The dynamic range of useful temporal fine structure cues for speech in the presence of a competing talker. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2011; 130:2162-72. [PMID: 21973370 DOI: 10.1121/1.3625237] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
Within an auditory channel, the speech waveform contains both temporal envelope (E(O)) and temporal fine structure (TFS) information. Vocoder processing extracts a modified version of the temporal envelope (E') within each channel and uses it to modulate a channel carrier. The resulting signal, E'(Carr), has reduced information content compared to the original "E(O) + TFS" signal. The dynamic range over which listeners make additional use of E(O) + TFS over E'(Carr) cues was investigated in a competing-speech task. The target-and-background mixture was processed using a 30-channel vocoder. In each channel, E(O) + TFS replaced E'(Carr) at either the peaks or the valleys of the signal. The replacement decision was based on comparing the short-term channel level to a parametrically varied "switching threshold," expressed relative to the long-term channel level. Intelligibility was measured as a function of switching threshold, carrier type, target-to-background ratio, and replacement method. Scores showed a dependence on all four parameters. Derived intensity-importance functions (IIFs) showed that E(O) + TFS information from 8-13 dB below to 10 dB above the channel long-term level was important. When E(O) + TFS information was added at the peaks, IIFs peaked around -2 dB, but when E(O) + TFS information was added at the valleys, the peaks lay around +1 dB.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom.
| | | | | |
Collapse
|
20
|
Stone MA, Füllgrabe C, Moore BCJ. Relative contribution to speech intelligibility of different envelope modulation rates within the speech dynamic range. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 128:2127-2137. [PMID: 20968383 DOI: 10.1121/1.3479546] [Citation(s) in RCA: 22] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
The contribution of envelope cues at different rates to intelligibility in a competing-speech task was measured as a function of the short-term envelope level. The target and background mixture was processed using tone vocoders. Envelope signals for each vocoder channel were simultaneously extracted with two low-pass filters, the cutoff frequency of one filter (L) being two octaves below that of the other (H). The envelope from the H filter was used at the peaks and that from the L filter at valleys, or vice versa. This was achieved by cross-fading between the two envelope signals based on a "switching threshold" that was parametrically varied relative to the long-term RMS level of the channel signal. When the cutoff frequencies of the H and L filters were 50 and 12.5 Hz, changes in speech intelligibility occurred mainly when the switching threshold was between -18 and +10 dB. The range was slightly narrower when the cutoff frequencies of the H and L filters were 200 and 50 Hz. Intensity-importance functions for higher-rate envelope modulations suggested that levels ranging from 20 dB below to about 10 dB above the channel RMS level were important, with maximum importance for levels around -5 dB.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, England.
| | | | | |
Collapse
|
21
|
Interactions between unsupervised learning and the degree of spectral mismatch on short-term perceptual adaptation to spectrally shifted speech. Ear Hear 2010; 30:238-49. [PMID: 19194293 DOI: 10.1097/aud.0b013e31819769ac] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Cochlear implant listeners are able to at least partially adapt to the spectral mismatch associated with the implant device and speech processor via daily exposure and/or explicit training. The overall goal of this study was to investigate interactions between short-term unsupervised learning (i.e., passive adaptation) and the degree of spectral mismatch in normal-hearing listeners' adaptation to spectrally shifted vowels. DESIGN Normal-hearing subjects were tested while listening to acoustic cochlear implant simulations. Unsupervised learning was measured by testing vowel recognition repeatedly over a 5 day period; no feedback or explicit training was provided. In experiment 1, subjects listened to 8-channel, sine-wave vocoded speech. The spectral envelope was compressed to simulate a 16 mm cochlear implant electrode array. The analysis bands were fixed and the compressed spectral envelope was linearly shifted toward the base by 3.6, 6, or 8.3 mm to simulate different insertion depths of the electrode array, resulting in a slight, moderate, or severe spectral shift. In experiment 2, half the subjects were exclusively exposed to a severe shift with 8 or 16 channels (exclusive groups), and half the subjects were exposed to 8-channel severely shifted speech, 16-channel severely shifted speech, and 8-channel moderately shifted speech, alternately presented within each test session (mixed group). The region of stimulation in the cochlea was fixed (16 mm in extent and 15 mm from the apex) and the analysis bands were manipulated to create the spectral shift conditions. To determine whether increased spectral resolution would improve adaptation, subjects were exposed to 8- or 16-channel severely shifted speech. RESULTS In experiment 1, at the end of the adaptation period, there was no significant difference between 8-channel speech that was spectrally matched and that shifted by 3.6 mm. There was a significant, but less-complete, adaptation to the 6 mm shift and no adaptation to the 8.3 mm shift. In experiment 2, for the mixed exposure group, there was significant adaptation to severely shifted speech with 8 channels and even greater adaptation with 16 channels. For the exclusive exposure group, there was no significant adaptation to severely shifted speech with either 8 or 16 channels. CONCLUSIONS These findings suggest that listeners are able to passively adapt to spectral shifts up to 6 mm. For spectral shifts beyond 6 mm, some passive adaptation was observed with mixed exposure to a smaller spectral shift, even at the expense of some low frequency information. Mixed exposure to the smaller shift may have enhanced listeners' access to spectral envelope details that were not accessible when listening exclusively to severely shifted speech. The results suggest that the range of spectral mismatch that can support passive adaptation may be larger than previously reported. Some amount of passive adaptation may be possible with severely shifted speech by exposing listeners to a relatively small mismatch in conjunction with the severe mismatch.
Collapse
|
22
|
Siciliano CM, Faulkner A, Rosen S, Mair K. Resistance to learning binaurally mismatched frequency-to-place maps: implications for bilateral stimulation with cochlear implants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 127:1645-60. [PMID: 20329863 DOI: 10.1121/1.3293002] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
Simulations of monaural cochlear implants in normal hearing listeners have shown that the deleterious effects of upward spectral shifting on speech perception can be overcome with training. This study simulates bilateral stimulation with a unilateral spectral shift to investigate whether listeners can adapt to upward-shifted speech information presented together with contralateral unshifted information. A six-channel, dichotic, interleaved sine-carrier vocoder simulated a binaurally mismatched frequency-to-place map. Odd channels were presented to one ear with an upward frequency shift equivalent to 6 mm on the basilar membrane, while even channels were presented to the contralateral ear unshifted. In Experiment 1, listeners were trained for 5.3 h with either the binaurally mismatched processor or with just the shifted monaural bands. In Experiment 2, the duration of training was 10 h, and the trained condition alternated between those of Experiment 1. While listeners showed learning in both experiments, intelligibility with the binaurally mismatched processor never exceeded, intelligibility with just the three unshifted bands, suggesting that listeners did not benefit from combining the mismatched maps, even though there was clear scope to do so. Frequency-place map alignment may thus be of importance when optimizing bilateral devices of the type studied here.
Collapse
Affiliation(s)
- Catherine M Siciliano
- Speech, Hearing and Phonetic Sciences, Division of Psychology and Language Sciences, UCL, Chandler House, 2 Wakefield Street, London WC1N 1PF, United Kingdom.
| | | | | | | |
Collapse
|
23
|
Stone MA, Füllgrabe C, Moore BCJ. Benefit of high-rate envelope cues in vocoder processing: effect of number of channels and spectral region. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 124:2272-82. [PMID: 19062865 DOI: 10.1121/1.2968678] [Citation(s) in RCA: 38] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
In cochlear implants, or vocoder simulations of cochlear implants, the transmission of envelope cues at high rates (related to voice fundamental frequency, f0) may be limited by the widths of the filters used to form the channels and/or by the cutoff frequency, f(lp), of the low-pass filters used for envelope extraction. The effect of varying f(lp) in tone and noise vocoders was investigated for channel numbers, N, from 6 to 18. As N increased, the widths of the channels decreased. The value of f(lp) was 45 Hz (envelope or "E" filter), or 180 Hz (pitch or "P" filter). The following combinations of cutoff frequencies were used for channels below and above 1500 Hz, respectively: EE, PE, EP, and PP. Results from a competing-talker task showed that the tone vocoder led to better intelligibility than the noise vocoder. The PP condition led to the best intelligibility and the EE condition to the worst. For N=6, intelligibility was better for condition PE than for condition EP. For N=18, the reverse was true. The results indicate that the channel bandwidths can compromise the transmission of f0-related envelope information, and suggest that vocoder simulations of cochlear-implant processing have limitations.
Collapse
Affiliation(s)
- Michael A Stone
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom.
| | | | | |
Collapse
|
24
|
Goupell MJ, Laback B, Majdak P, Baumgartner WD. Effects of upper-frequency boundary and spectral warping on speech intelligibility in electrical stimulation. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2008; 123:2295-309. [PMID: 18397034 PMCID: PMC3061454 DOI: 10.1121/1.2831738] [Citation(s) in RCA: 12] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Speech understanding was tested for seven listeners using 12-electrode Med-El cochlear implants (CIs) and six normal-hearing listeners using a CI simulation. Eighteen different types of processing were evaluated, which varied the frequency-to-tonotopic place mapping and the upper boundary of the frequency and stimulation range. Spectrally unwarped and warped conditions were included. Unlike previous studies on this topic, the lower boundary of the frequency and stimulation range was fixed while the upper boundary was varied. For the unwarped conditions, only eight to ten channels were needed in both quiet and noise to achieve no significant degradation in speech understanding compared to the normal 12-electrode speech processing. The unwarped conditions were often the best conditions for understanding speech; however, small changes in frequency-to-place mapping (<0.77 octaves for the most basal electrode) yielded no significant degradation in performance from the nearest unwarped condition. A second experiment measured the effect of feedback training for both the unwarped and warped conditions. Improvements were found for the unwarped and frequency-expanded conditions, but not for the compressed condition. These results have implications for new CI processing strategies, such as the inclusion of spectral localization cues.
Collapse
Affiliation(s)
- Matthew J Goupell
- Acoustics Research Institute, Austrian Academy of Sciences, Wohllebengasse 12-14, A-1040 Vienna, Austria.
| | | | | | | |
Collapse
|
25
|
Fu QJ, Galvin JJ. Maximizing cochlear implant patients' performance with advanced speech training procedures. Hear Res 2007; 242:198-208. [PMID: 18295992 DOI: 10.1016/j.heares.2007.11.010] [Citation(s) in RCA: 89] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/04/2007] [Revised: 11/21/2007] [Accepted: 11/22/2007] [Indexed: 11/19/2022]
Abstract
Advances in implant technology and speech processing have provided great benefit to many cochlear implant patients. However, some patients receive little benefit from the latest technology, even after many years' experience with the device. Moreover, even the best cochlear implant performers have great difficulty understanding speech in background noise, and music perception and appreciation remain major challenges. Recent studies have shown that targeted auditory training can significantly improve cochlear implant patients' speech recognition performance. Such benefits are not only observed in poorly performing patients, but also in good performers under difficult listening conditions (e.g., speech noise, telephone speech, music, etc.). Targeted auditory training has also been shown to enhance performance gains provided by new implant devices and/or speech processing strategies. These studies suggest that cochlear implantation alone may not fully meet the needs of many patients, and that additional auditory rehabilitation may be needed to maximize the benefits of the implant device. Continuing research will aid in the development of efficient and effective training protocols and materials, thereby minimizing the costs (in terms of time, effort and resources) associated with auditory rehabilitation while maximizing the benefits of cochlear implantation for all recipients.
Collapse
Affiliation(s)
- Qian-Jie Fu
- Department of Auditory Implants and Perception, House Ear Institute, 2100 West Third Street, Los Angeles, CA 90057, USA.
| | | |
Collapse
|
26
|
Faulkner A. Adaptation to distorted frequency-to-place maps: implications of simulations in normal listeners for cochlear implants and electroacoustic stimulation. Audiol Neurootol 2006; 11 Suppl 1:21-6. [PMID: 17063007 DOI: 10.1159/000095610] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2006] [Accepted: 06/12/2006] [Indexed: 11/19/2022] Open
Abstract
The ideal cochlear implant electrode array positioning enables stimulation over a range of cochlear positions whose characteristic frequencies cover the frequency range of speech and match the speech processor filter frequencies. However, the electrode positions achieved in practice may not meet this specification. Users of conventional monaural cochlear implants seem able to perceptually adapt to a mismatch of speech processor filters to electrode positions. In electroacoustic stimulation, it is important to consider possible inconsistencies between acoustic and electrical frequency-to-place mapping. Two simulation studies are outlined that address normal listeners' ability to perceive speech presented through distorted frequency maps. The first presented a map that is spectrally warped around a 10-mm medial cochlear area. Listeners were able to adapt to this map after a few hours of training. The second study presented a binaural mapping in which one ear was subject to a 6-mm basalward shift. Here listeners were unable to learn to integrate speech information across the two mismatched ears, rather they seem to learn to ignore the shifted information. Frequency-to-place mapping is likely to be an important factor in the successful use of a combination of electrical and acoustic hearing.
Collapse
|