1
|
Cychosz M, Winn MB, Goupell MJ. How to vocode: Using channel vocoders for cochlear-implant research. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2407-2437. [PMID: 38568143 PMCID: PMC10994674 DOI: 10.1121/10.0025274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Revised: 02/14/2024] [Accepted: 02/23/2024] [Indexed: 04/05/2024]
Abstract
The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, California 90095, USA
| | - Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, College Park, Maryland 20742, USA
| |
Collapse
|
2
|
Sagi E, Azadpour M, Neukam J, Capach NH, Svirsky MA. Reducing interaural tonotopic mismatch preserves binaural unmasking in cochlear implant simulations of single-sided deafness. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:2316. [PMID: 34717490 PMCID: PMC8637719 DOI: 10.1121/10.0006446] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/20/2021] [Revised: 09/02/2021] [Accepted: 09/08/2021] [Indexed: 06/13/2023]
Abstract
Binaural unmasking, a key feature of normal binaural hearing, can refer to the improved intelligibility of masked speech by adding masking that facilitates perceived separation of target and masker. A question relevant for cochlear implant users with single-sided deafness (SSD-CI) is whether binaural unmasking can still be achieved if the additional masking is spectrally degraded and shifted. CIs restore some aspects of binaural hearing to these listeners, although binaural unmasking remains limited. Notably, these listeners may experience a mismatch between the frequency information perceived through the CI and that perceived by their normal hearing ear. Employing acoustic simulations of SSD-CI with normal hearing listeners, the present study confirms a previous simulation study that binaural unmasking is severely limited when interaural frequency mismatch between the input frequency range and simulated place of stimulation exceeds 1-2 mm. The present study also shows that binaural unmasking is largely retained when the input frequency range is adjusted to match simulated place of stimulation, even at the expense of removing low-frequency information. This result bears implications for the mechanisms driving the type of binaural unmasking of the present study and for mapping the frequency range of the CI speech processor in SSD-CI users.
Collapse
Affiliation(s)
- Elad Sagi
- Department of Otolaryngology-Head & Neck Surgery, New York University Grossman School of Medicine, 550 First Avenue, New York, New York 10016, USA
| | - Mahan Azadpour
- Department of Otolaryngology-Head & Neck Surgery, New York University Grossman School of Medicine, 550 First Avenue, New York, New York 10016, USA
| | - Jonathan Neukam
- Department of Otolaryngology-Head & Neck Surgery, New York University Grossman School of Medicine, 550 First Avenue, New York, New York 10016, USA
| | - Nicole Hope Capach
- Department of Otolaryngology-Head & Neck Surgery, New York University Grossman School of Medicine, 550 First Avenue, New York, New York 10016, USA
| | - Mario A Svirsky
- Department of Otolaryngology-Head & Neck Surgery, New York University Grossman School of Medicine, 550 First Avenue, New York, New York 10016, USA
| |
Collapse
|
3
|
The effect of increased channel interaction on speech perception with cochlear implants. Sci Rep 2021; 11:10383. [PMID: 34001987 PMCID: PMC8128897 DOI: 10.1038/s41598-021-89932-8] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2021] [Accepted: 04/29/2021] [Indexed: 11/30/2022] Open
Abstract
Cochlear implants (CIs) are neuroprostheses that partially restore hearing for people with severe-to-profound hearing loss. While CIs can provide good speech perception in quiet listening situations for many, they fail to do so in environments with interfering sounds for most listeners. Previous research suggests that this is due to detrimental interaction effects between CI electrode channels, limiting their function to convey frequency-specific information, but evidence is still scarce. In this study, an experimental manipulation called spectral blurring was used to increase channel interaction in CI listeners using Advanced Bionics devices with HiFocus 1J and MS electrode arrays to directly investigate its causal effect on speech perception. Instead of using a single electrode per channel as in standard CI processing, spectral blurring used up to 6 electrodes per channel simultaneously to increase the overlap between adjacent frequency channels as would occur in cases with severe channel interaction. Results demonstrated that this manipulation significantly degraded CI speech perception in quiet by 15% and speech reception thresholds in babble noise by 5 dB when all channels were blurred by a factor of 6. Importantly, when channel interaction was increased just on a subset of electrodes, speech scores were mostly unaffected and were only significantly degraded when the 5 most apical channels were blurred. These apical channels convey information up to 1 kHz at the apical end of the electrode array and are typically located at angular insertion depths of about 250 up to 500°. These results confirm and extend earlier findings indicating that CI speech perception may not benefit from deactivating individual channels along the array and that efforts should instead be directed towards reducing channel interaction per se and in particular for the most-apical electrodes. Hereby, causal methods such as spectral blurring could be used in future research to control channel interaction effects within listeners for evaluating compensation strategies.
Collapse
|
4
|
Chavant M, Hervais-Adelman A, Macherey O. Perceptual Learning of Vocoded Speech With and Without Contralateral Hearing: Implications for Cochlear Implant Rehabilitation. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:196-205. [PMID: 33267729 DOI: 10.1044/2020_jslhr-20-00385] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Purpose An increasing number of individuals with residual or even normal contralateral hearing are being considered for cochlear implantation. It remains unknown whether the presence of contralateral hearing is beneficial or detrimental to their perceptual learning of cochlear implant (CI)-processed speech. The aim of this experiment was to provide a first insight into this question using acoustic simulations of CI processing. Method Sixty normal-hearing listeners took part in an auditory perceptual learning experiment. Each subject was randomly assigned to one of three groups of 20 referred to as NORMAL, LOWPASS, and NOTHING. The experiment consisted of two test phases separated by a training phase. In the test phases, all subjects were tested on recognition of monosyllabic words passed through a six-channel "PSHC" vocoder presented to a single ear. In the training phase, which consisted of listening to a 25-min audio book, all subjects were also presented with the same vocoded speech in one ear but the signal they received in their other ear differed across groups. The NORMAL group was presented with the unprocessed speech signal, the LOWPASS group with a low-pass filtered version of the speech signal, and the NOTHING group with no sound at all. Results The improvement in speech scores following training was significantly smaller for the NORMAL than for the LOWPASS and NOTHING groups. Conclusions This study suggests that the presentation of normal speech in the contralateral ear reduces or slows down perceptual learning of vocoded speech but that an unintelligible low-pass filtered contralateral signal does not have this effect. Potential implications for the rehabilitation of CI patients with partial or full contralateral hearing are discussed.
Collapse
Affiliation(s)
- Martin Chavant
- Aix-Marseille University, Centre National de la Recherche Scientifique, Centrale Marseille, Laboratoire de Mécanique et d'Acoustique, France
| | | | - Olivier Macherey
- Aix-Marseille University, Centre National de la Recherche Scientifique, Centrale Marseille, Laboratoire de Mécanique et d'Acoustique, France
| |
Collapse
|
5
|
The sound sensation of a pure tone in cochlear implant recipients with single-sided deafness. PLoS One 2020; 15:e0235504. [PMID: 32658911 PMCID: PMC7357770 DOI: 10.1371/journal.pone.0235504] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2020] [Accepted: 06/17/2020] [Indexed: 11/19/2022] Open
Abstract
Ten cochlear implant (CI) users with single-sided deafness were asked to vary the parameters of an acoustic sound played to their contralateral ear to characterize the perception evoked by a pure tone played through the direct audio input of their CI. Two frequencies, centered on an apical and a medial electrode, were tested. In six subjects, the electrode positions were estimated on CT scans. The study was divided in 3 experiments in which the parameters of the acoustic sound varied. The listeners had to vary the frequency of a pure tone (Exp.1), the center frequency and the bandwidth of a filter applied to a harmonic complex sound (Exp.2), and the frequency of the components and the inharmonicity factor of a complex sound (Exp.3). Two testing sessions were performed at 3 and 12 months after activation. The mean results of Exp. 1 showed that the frequency of the matched tone was significantly lower for the apical than for the medial stimulus. In Exp.2, the mean center frequencies of the filters were also significantly lower for the apical than for the medial stimulus. As this parameter modifies the energy ratio between the high and low-frequency components, this result suggests that the medial stimulus was perceived with a brighter timbre than the apical stimulus. In Exp.3, the mean frequencies of the components were not significantly different between the sounds resulting from the stimulation of the two electrodes, but were significantly lower at the12-month session compared to the 3-month visit. These results suggest that a change in place of excitation may be perceived as a change in timbre rather than a change in pitch, and that an effect of adaptation can be observed.
Collapse
|
6
|
Karoui C, James C, Barone P, Bakhos D, Marx M, Macherey O. Searching for the Sound of a Cochlear Implant: Evaluation of Different Vocoder Parameters by Cochlear Implant Users With Single-Sided Deafness. Trends Hear 2020; 23:2331216519866029. [PMID: 31533581 PMCID: PMC6753516 DOI: 10.1177/2331216519866029] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Cochlear implantation in subjects with single-sided deafness (SSD) offers a unique opportunity to directly compare the percepts evoked by a cochlear implant (CI) with those evoked acoustically. Here, nine SSD-CI users performed a forced-choice task evaluating the similarity of speech processed by their CI with speech processed by several vocoders presented to their healthy ear. In each trial, subjects heard two intervals: their CI followed by a certain vocoder in Interval 1 and their CI followed by a different vocoder in Interval 2. The vocoders differed either (i) in carrier type-(sinusoidal [SINE], bandfiltered noise [NOISE], and pulse-spreading harmonic complex) or (ii) in frequency mismatch between the analysis and synthesis frequency ranges-(no mismatch, and two frequency-mismatched conditions of 2 and 4 equivalent rectangular bandwidths [ERBs]). Subjects had to state in which of the two intervals the CI and vocoder sounds were more similar. Despite a large intersubject variability, the PSHC vocoder was judged significantly more similar to the CI than SINE or NOISE vocoders. Furthermore, the No-mismatch and 2-ERB mismatch vocoders were judged significantly more similar to the CI than the 4-ERB mismatch vocoder. The mismatch data were also interpreted by comparing spiral ganglion characteristic frequencies with electrode contact positions determined from postoperative computed tomography scans. Only one subject demonstrated a pattern of preference consistent with adaptation to the CI sound processor frequency-to-electrode allocation table and two subjects showed possible partial adaptation. Those subjects with adaptation patterns presented overall small and consistent frequency mismatches across their electrode arrays.
Collapse
Affiliation(s)
- Chadlia Karoui
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Cochlear France SAS, Toulouse, France
| | - Chris James
- Cochlear France SAS, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Pascal Barone
- Centre de Recherche Cerveau et Cognition, Toulouse, France
| | - David Bakhos
- Université François-Rabelais de Tours, CHRU de Tours, France.,Ear Nose and Throat department, CHUR de Tours, Tours, France
| | - Mathieu Marx
- Centre de Recherche Cerveau et Cognition, Toulouse, France.,Department of Otology-Neurotology and Skull Base Surgery, Purpan University Hospital, Toulouse, France
| | - Olivier Macherey
- Aix Marseille University, CNRS, Centrale Marseille, LMA, Marseille, France
| |
Collapse
|
7
|
Wess JM, Spencer NJ, Bernstein JGW. Counting or discriminating the number of voices to assess binaural fusion with single-sided vocoders. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:446. [PMID: 32006956 PMCID: PMC7043860 DOI: 10.1121/10.0000511] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/20/2019] [Revised: 11/27/2019] [Accepted: 12/06/2019] [Indexed: 06/10/2023]
Abstract
For single-sided deafness cochlear-implant (SSD-CI) listeners, different peripheral representations for electric versus acoustic stimulation, combined with interaural frequency mismatch, might limit the ability to perceive bilaterally presented speech as a single voice. The assessment of binaural fusion often relies on subjective report, which requires listeners to have some understanding of the perceptual phenomenon of object formation. Two experiments explored whether binaural fusion could instead be assessed using judgments of the number of voices in a mixture. In an SSD-CI simulation, normal-hearing listeners were presented with one or two "diotic" voices (i.e., unprocessed in one ear and noise-vocoded in the other) in a mixture with additional monaural voices. In experiment 1, listeners reported how many voices they heard. Listeners generally counted the diotic speech as two separate voices, regardless of interaural frequency mismatch. In experiment 2, listeners identified which of two mixtures contained diotic speech. Listeners performed significantly better with interaurally frequency-matched than with frequency-mismatched stimuli. These contrasting results suggest that listeners experienced partial fusion: not enough to count the diotic speech as one voice, but enough to detect its presence. The diotic-speech detection task (experiment 2) might provide a tool to evaluate fusion and optimize frequency mapping for SSD-CI patients.
Collapse
Affiliation(s)
- Jessica M Wess
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, Maryland 20889, USA
| | - Nathaniel J Spencer
- Air Force Research Laboratory, Wright Patterson Air Force Base, Ohio 45433, USA
| | - Joshua G W Bernstein
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, Maryland 20889, USA
| |
Collapse
|
8
|
Adel Y, Nagel S, Weissgerber T, Baumann U, Macherey O. Pitch Matching in Cochlear Implant Users With Single-Sided Deafness: Effects of Electrode Position and Acoustic Stimulus Type. Front Neurosci 2019; 13:1119. [PMID: 31736684 PMCID: PMC6839387 DOI: 10.3389/fnins.2019.01119] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2019] [Accepted: 10/03/2019] [Indexed: 01/27/2023] Open
Abstract
Previous studies in patients with single-sided deafness (SSD) have reported results of pitch comparisons between electric stimulation of their cochlear implant (CI) and acoustic stimulation presented to their near-normal hearing contralateral ear. These comparisons typically used sinusoids, although the percept elicited by electric stimulation may be closer to a wideband stimulus. Furthermore, it has been shown that pitch comparisons between sounds with different timbres is a difficult task and subjected to various types of range biases. The present study aims to introduce a method to minimize non-sensory biases, and to investigate the effect of different acoustic stimulus types on the frequency and variability of the electric-acoustic pitch matches. Pitch matches were collected from 13 CI users with SSD using the binary search procedure. Electric stimulation was presented at either an apical or a middle electrode position, at a rate of 800 pps. Acoustic stimulus types were sinusoids (SINE), 1/3-octave wide narrow bands of Gaussian noises (NBN), or 1/3-octave wide pulse spreading harmonic complexes (PSHC). On the one hand, NBN and PSHC are presumed to better mimic the spread of excitation produced by a single-electrode stimulation than SINE. On the other hand, SINE and PSHC contain less inherent fluctuations than NBN and may therefore provide a temporal pattern closer to that produced by a constant-amplitude electric pulse train. Analysis of mean pitch match variance showed no differences between stimulus types. However, mean pitch matches showed effects of electrode position and stimulus type, with the middle electrode always matched to a higher frequency than the apical one (p < 0.001), and significantly higher across-subject pitch matches for PSHC compared with SINE (p = 0.017). Mean pitch matches for all stimulus types were better predicted by place-dependent characteristic frequencies (CFs) based on an organ of Corti map compared with a spiral ganglion map. CF predictions were closest to pitch matches with SINE for the apical electrode position, and conversely with NBN or PSHC for the middle electrode position. These results provide evidence that the choice of acoustic stimulus type can have a significant effect on electric-acoustic pitch matching.
Collapse
Affiliation(s)
- Youssef Adel
- Audiological Acoustics, Department of Otorhinolaryngology, University Hospital Frankfurt, Frankfurt, Germany
| | - Sharon Nagel
- Audiological Acoustics, Department of Otorhinolaryngology, University Hospital Frankfurt, Frankfurt, Germany
| | - Tobias Weissgerber
- Audiological Acoustics, Department of Otorhinolaryngology, University Hospital Frankfurt, Frankfurt, Germany
| | - Uwe Baumann
- Audiological Acoustics, Department of Otorhinolaryngology, University Hospital Frankfurt, Frankfurt, Germany
| | - Olivier Macherey
- Aix-Marseille University, CNRS, Centrale Marseille, LMA, Marseille, France
| |
Collapse
|
9
|
Steinmetzger K, Rosen S. The role of envelope periodicity in the perception of masked speech with simulated and real cochlear implants. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:885. [PMID: 30180719 DOI: 10.1121/1.5049584] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/04/2018] [Accepted: 07/22/2018] [Indexed: 06/08/2023]
Abstract
In normal hearing, complex tones with pitch-related periodic envelope modulations are far less effective maskers of speech than aperiodic noise. Here, it is shown that this masker-periodicity benefit is diminished in noise-vocoder simulations of cochlear implants (CIs) and further reduced with real CIs. Nevertheless, both listener groups still benefitted significantly from masker periodicity, despite the lack of salient spectral pitch cues. The main reason for the smaller effect observed in CI users is thought to be an even stronger channel interaction than in the CI simulations, which smears out the random envelope modulations that are characteristic for aperiodic sounds. In contrast, neither interferers that were amplitude-modulated at a rate of 10 Hz nor maskers with envelopes specifically designed to reveal the target speech enabled a masking release in CI users. Hence, even at the high signal-to-noise ratios at which they were tested, CI users can still exploit pitch cues transmitted by the temporal envelope of a non-speech masker, whereas slow amplitude modulations of the masker envelope are no longer helpful.
Collapse
Affiliation(s)
- Kurt Steinmetzger
- Speech, Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London, WC1N 1PF, United Kingdom
| | - Stuart Rosen
- Speech, Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London, WC1N 1PF, United Kingdom
| |
Collapse
|