1
|
Levi SV, Cheng HS, O'Brien G, Harel D. A Picture May Be Worth 1,000 Words, but Is It Worth a Letter? Examining Whether the Choice of Label Affects the Perception of Speech Sounds. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:2115-2127. [PMID: 38754023 DOI: 10.1044/2024_jslhr-23-00616] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/18/2024]
Abstract
PURPOSE Researchers often use identification or goodness rating tasks to assess speech perception for different populations. These tasks provide useful information about a listener's willingness to accept a range of acoustically variable stimuli as belonging to the same category and also about assessing how stimuli that are labeled the same may not be perceived as equally good versions of a particular speech sound. Many methodological aspects of these simple tasks have been tested, but one aspect that has not is the choice of label. In this study, we examine response patterns to images versus letters, as studies with different populations (children vs. adults) or different methods (typical behavioral study vs. visual world paradigm) may vary in the type of label used. METHOD Eighty-one adult listeners completed phoneme identification and goodness ratings tasks with either images of response options (a picture of a bear and a picture of a pear) or with letter labels (a capital B and P). RESULTS The results suggest that choice of label does not alter performance within the tasks studied here. In addition, the results did show the expected finding that the slope of the response curve is steeper in an identification task than in a goodness rating task. CONCLUSION These results suggest that it is possible to compare across studies that use different response options, a benefit to research and practice because letter labels can be used for nonimageable words and nonwords, whereas images may be best used for participants who are younger or have poorer reading skills.
Collapse
Affiliation(s)
- Susannah V Levi
- Department of Communicative Science and Disorders, New York University, NY
| | - Hung-Shao Cheng
- Department of Communicative Science and Disorders, New York University, NY
| | | | - Daphna Harel
- Department of Applied Statistics, Social Sciences, and Humanities, New York University, NY
| |
Collapse
|
2
|
Cychosz M, Winn MB, Goupell MJ. How to vocode: Using channel vocoders for cochlear-implant research. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2407-2437. [PMID: 38568143 PMCID: PMC10994674 DOI: 10.1121/10.0025274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Accepted: 02/23/2024] [Indexed: 04/05/2024]
Abstract
The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, California 90095, USA
| | - Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, College Park, Maryland 20742, USA
| |
Collapse
|
3
|
Bosen AK, Doria GM. Identifying Links Between Latent Memory and Speech Recognition Factors. Ear Hear 2024; 45:351-369. [PMID: 37882100 PMCID: PMC10922378 DOI: 10.1097/aud.0000000000001430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2023]
Abstract
OBJECTIVES The link between memory ability and speech recognition accuracy is often examined by correlating summary measures of performance across various tasks, but interpretation of such correlations critically depends on assumptions about how these measures map onto underlying factors of interest. The present work presents an alternative approach, wherein latent factor models are fit to trial-level data from multiple tasks to directly test hypotheses about the underlying structure of memory and the extent to which latent memory factors are associated with individual differences in speech recognition accuracy. Latent factor models with different numbers of factors were fit to the data and compared to one another to select the structures which best explained vocoded sentence recognition in a two-talker masker across a range of target-to-masker ratios, performance on three memory tasks, and the link between sentence recognition and memory. DESIGN Young adults with normal hearing (N = 52 for the memory tasks, of which 21 participants also completed the sentence recognition task) completed three memory tasks and one sentence recognition task: reading span, auditory digit span, visual free recall of words, and recognition of 16-channel vocoded Perceptually Robust English Sentence Test Open-set sentences in the presence of a two-talker masker at target-to-masker ratios between +10 and 0 dB. Correlations between summary measures of memory task performance and sentence recognition accuracy were calculated for comparison to prior work, and latent factor models were fit to trial-level data and compared against one another to identify the number of latent factors which best explains the data. Models with one or two latent factors were fit to the sentence recognition data and models with one, two, or three latent factors were fit to the memory task data. Based on findings with these models, full models that linked one speech factor to one, two, or three memory factors were fit to the full data set. Models were compared via Expected Log pointwise Predictive Density and post hoc inspection of model parameters. RESULTS Summary measures were positively correlated across memory tasks and sentence recognition. Latent factor models revealed that sentence recognition accuracy was best explained by a single factor that varied across participants. Memory task performance was best explained by two latent factors, of which one was generally associated with performance on all three tasks and the other was specific to digit span recall accuracy at lists of six digits or more. When these models were combined, the general memory factor was closely related to the sentence recognition factor, whereas the factor specific to digit span had no apparent association with sentence recognition. CONCLUSIONS Comparison of latent factor models enables testing hypotheses about the underlying structure linking cognition and speech recognition. This approach showed that multiple memory tasks assess a common latent factor that is related to individual differences in sentence recognition, although performance on some tasks was associated with multiple factors. Thus, while these tasks provide some convergent assessment of common latent factors, caution is needed when interpreting what they tell us about speech recognition.
Collapse
|
4
|
Noble AR, Halverson DM, Resnick J, Broncheau M, Rubinstein JT, Horn DL. Spectral Resolution and Speech Perception in Cochlear Implanted School-Aged Children. Otolaryngol Head Neck Surg 2024; 170:230-238. [PMID: 37365946 PMCID: PMC10836047 DOI: 10.1002/ohn.408] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2023] [Revised: 05/03/2023] [Accepted: 06/04/2023] [Indexed: 06/28/2023]
Abstract
OBJECTIVE Cochlear implantation of prelingually deaf infants provides auditory input sufficient to develop spoken language; however, outcomes remain variable. Inability to participate in speech perception testing limits testing device efficacy in young listeners. In postlingually implanted adults (aCI), speech perception correlates with spectral resolution an ability that relies independently on frequency resolution (FR) and spectral modulation sensitivity (SMS). The correlation of spectral resolution to speech perception is unknown in prelingually implanted children (cCI). In this study, FR and SMS were measured using a spectral ripple discrimination (SRD) task and were correlated with vowel and consonant identification. It was hypothesized that prelingually deaf cCI would show immature SMS relative to postlingually deaf aCI and that FR would correlate with speech identification. STUDY DESIGN Cross-sectional study. SETTING In-person, booth testing. METHODS SRD was used to determine the highest spectral ripple density perceived at various modulation depths. FR and SMS were derived from spectral modulation transfer functions. Vowel and consonant identification was measured; SRD performance and speech identification were analyzed for correlation. RESULTS Fifteen prelingually implanted cCI and 13 postlingually implanted aCI were included. FR and SMS were similar between cCI and aCI. Better FR was associated with better speech identification for most measures. CONCLUSION Prelingually implanted cCI demonstrated adult-like FR and SMS; additionally, FR correlated with speech identification. FR may be a measure of CI efficacy in young listeners.
Collapse
Affiliation(s)
- Anisha R. Noble
- Division of Pediatric Otolaryngology – Head and Neck Surgery, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH, USA
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology – Head and Neck Surgery, University of Washington, Seattle, WA, USA
| | - Destinee M. Halverson
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology – Head and Neck Surgery, University of Washington, Seattle, WA, USA
| | - Jesse Resnick
- Department of Internal Medicine, University of Michigan, Ann Arbor, MI, USA
| | - Mariette Broncheau
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology – Head and Neck Surgery, University of Washington, Seattle, WA, USA
| | - Jay T. Rubinstein
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology – Head and Neck Surgery, University of Washington, Seattle, WA, USA
| | - David L. Horn
- Virginia Merrill Bloedel Hearing Research Center, Department of Otolaryngology – Head and Neck Surgery, University of Washington, Seattle, WA, USA
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| |
Collapse
|
5
|
Cychosz M, Xu K, Fu QJ. Effects of spectral smearing on speech understanding and masking release in simulated bilateral cochlear implants. PLoS One 2023; 18:e0287728. [PMID: 37917727 PMCID: PMC10621938 DOI: 10.1371/journal.pone.0287728] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Accepted: 06/11/2023] [Indexed: 11/04/2023] Open
Abstract
Differences in spectro-temporal degradation may explain some variability in cochlear implant users' speech outcomes. The present study employs vocoder simulations on listeners with typical hearing to evaluate how differences in degree of channel interaction across ears affects spatial speech recognition. Speech recognition thresholds and spatial release from masking were measured in 16 normal-hearing subjects listening to simulated bilateral cochlear implants. 16-channel sine-vocoded speech simulated limited, broad, or mixed channel interaction, in dichotic and diotic target-masker conditions, across ears. Thresholds were highest with broad channel interaction in both ears but improved when interaction decreased in one ear and again in both ears. Masking release was apparent across conditions. Results from this simulation study on listeners with typical hearing show that channel interaction may impact speech recognition more than masking release, and may have implications for the effects of channel interaction on cochlear implant users' speech recognition outcomes.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, CA, United States of America
| | - Kevin Xu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States of America
| | - Qian-Jie Fu
- Department of Head and Neck Surgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States of America
| |
Collapse
|
6
|
de la Cruz-Pavía I, Eloy C, Perrineau-Hecklé P, Nazzi T, Cabrera L. Consonant bias in adult lexical processing under acoustically degraded listening conditions. JASA EXPRESS LETTERS 2023; 3:2892558. [PMID: 37220232 DOI: 10.1121/10.0019576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Accepted: 05/05/2023] [Indexed: 05/25/2023]
Abstract
Consonants facilitate lexical processing across many languages, including French. This study investigates whether acoustic degradation affects this phonological bias in an auditory lexical decision task. French words were processed using an eight-band vocoder, degrading their frequency modulations (FM) while preserving original amplitude modulations (AM). Adult French natives were presented with these French words, preceded by similarly processed pseudoword primes sharing their vowels, consonants, or neither. Results reveal a consonant bias in the listeners' accuracy and response times, despite the reduced spectral and FM information. These degraded conditions resemble current cochlear-implant processors, and attest to the robustness of this phonological bias.
Collapse
Affiliation(s)
- Irene de la Cruz-Pavía
- Department of Linguistics and Basque Studies, Universidad del País Vasco/Euskal Herriko Unibertsitatea, Vitoria-Gasteiz 01006, Spain
| | - Coraline Eloy
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Paula Perrineau-Hecklé
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Thierry Nazzi
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| | - Laurianne Cabrera
- Integrative Neuroscience and Cognition Center, Université Paris Cité, Centre National de la Recherche Scientifique, Paris 75006, , , , ,
| |
Collapse
|
7
|
Winn MB, Wright RA. Reconsidering commonly used stimuli in speech perception experiments. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:1394. [PMID: 36182291 DOI: 10.1121/10.0013415] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/24/2022] [Accepted: 07/18/2022] [Indexed: 06/16/2023]
Abstract
This paper examines some commonly used stimuli in speech perception experiments and raises questions about their use, or about the interpretations of previous results. The takeaway messages are: 1) the Hillenbrand vowels represent a particular dialect rather than a gold standard, and English vowels contain spectral dynamics that have been largely underappreciated, 2) the /ɑ/ context is very common but not clearly superior as a context for testing consonant perception, 3) /ɑ/ is particularly problematic when testing voice-onset-time perception because it introduces strong confounds in the formant transitions, 4) /dɑ/ is grossly overrepresented in neurophysiological studies and yet is insufficient as a generalized proxy for "speech perception," and 5) digit tests and matrix sentences including the coordinate response measure are systematically insensitive to important patterns in speech perception. Each of these stimulus sets and concepts is described with careful attention to their unique value and also cases where they might be misunderstood or over-interpreted.
Collapse
Affiliation(s)
- Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Richard A Wright
- Department of Linguistics, University of Washington, Seattle, Washington 98195, USA
| |
Collapse
|
8
|
Arjmandi MK, Jahn KN, Arenberg JG. Single-Channel Focused Thresholds Relate to Vowel Identification in Pediatric and Adult Cochlear Implant Listeners. Trends Hear 2022; 26:23312165221095364. [PMID: 35505617 PMCID: PMC9073113 DOI: 10.1177/23312165221095364] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Abstract
Speech recognition outcomes are highly variable among pediatric and adult cochlear implant (CI) listeners. Although there is some evidence that the quality of the electrode-neuron interface (ENI) contributes to this large variability in auditory perception, its relationship with speech outcomes is not well understood. Single-channel auditory detection thresholds measured in response to focused electrical fields (i.e., focused thresholds) are sensitive to properties of ENI quality, including electrode-neuron distance, intracochlear resistance, and neural health. In the present study, focused thresholds and speech perception abilities were assessed in 15 children and 21 adult CI listeners. Focused thresholds were measured for all active electrodes using a fast sweep procedure. Speech perception performance was evaluated by assessing listeners’ ability to identify vowels presented in /h-vowel-d/ context. Consistent with prior literature, focused thresholds were lower for children than for adults, but vowel identification did not differ significantly across age groups. Higher across-array average focused thresholds, which may indicate a relatively poor ENI quality, were associated with poorer vowel identification scores in both children and adults. Adult CI listeners with longer durations of deafness had higher focused thresholds. Findings from this study demonstrate that poor-quality ENIs may contribute to reduced speech outcomes for pediatric and adult CI listeners. Estimates of ENI quality (e.g., focused thresholds) may assist in developing customized programming interventions that serve to improve the transmission of spectral cues that are important in vowel identification.
Collapse
Affiliation(s)
- Meisam K Arjmandi
- Department of Otolaryngology - Head and Neck Surgery, 1811Harvard Medical School, Boston, MA, USA.,Eaton-Peabody Laboratories, 1866Massachusetts Eye and Ear, Boston, MA, USA.,Audiology Division, 1866Massachusetts Eye and Ear, Boston, MA, USA
| | - Kelly N Jahn
- Department of Otolaryngology - Head and Neck Surgery, 1811Harvard Medical School, Boston, MA, USA.,Eaton-Peabody Laboratories, 1866Massachusetts Eye and Ear, Boston, MA, USA.,Department of Speech, Language, and Hearing, University of Texas at Dallas, Richardson, TX, USA
| | - Julie G Arenberg
- Department of Otolaryngology - Head and Neck Surgery, 1811Harvard Medical School, Boston, MA, USA.,Eaton-Peabody Laboratories, 1866Massachusetts Eye and Ear, Boston, MA, USA.,Audiology Division, 1866Massachusetts Eye and Ear, Boston, MA, USA
| |
Collapse
|
9
|
Individual Variability in Recalibrating to Spectrally Shifted Speech: Implications for Cochlear Implants. Ear Hear 2021; 42:1412-1427. [PMID: 33795617 DOI: 10.1097/aud.0000000000001043] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Cochlear implant (CI) recipients are at a severe disadvantage compared with normal-hearing listeners in distinguishing consonants that differ by place of articulation because the key relevant spectral differences are degraded by the implant. One component of that degradation is the upward shifting of spectral energy that occurs with a shallow insertion depth of a CI. The present study aimed to systematically measure the effects of spectral shifting on word recognition and phoneme categorization by specifically controlling the amount of shifting and using stimuli whose identification specifically depends on perceiving frequency cues. We hypothesized that listeners would be biased toward perceiving phonemes that contain higher-frequency components because of the upward frequency shift and that intelligibility would decrease as spectral shifting increased. DESIGN Normal-hearing listeners (n = 15) heard sine wave-vocoded speech with simulated upward frequency shifts of 0, 2, 4, and 6 mm of cochlear space to simulate shallow CI insertion depth. Stimuli included monosyllabic words and /b/-/d/ and /∫/-/s/ continua that varied systematically by formant frequency transitions or frication noise spectral peaks, respectively. Recalibration to spectral shifting was operationally defined as shifting perceptual acoustic-phonetic mapping commensurate with the spectral shift. In other words, adjusting frequency expectations for both phonemes upward so that there is still a perceptual distinction, rather than hearing all upward-shifted phonemes as the higher-frequency member of the pair. RESULTS For moderate amounts of spectral shifting, group data suggested a general "halfway" recalibration to spectral shifting, but individual data suggested a notably different conclusion: half of the listeners were able to recalibrate fully, while the other halves of the listeners were utterly unable to categorize shifted speech with any reliability. There were no participants who demonstrated a pattern intermediate to these two extremes. Intelligibility of words decreased with greater amounts of spectral shifting, also showing loose clusters of better- and poorer-performing listeners. Phonetic analysis of word errors revealed certain cues were more susceptible to being compromised due to a frequency shift (place and manner of articulation), while voicing was robust to spectral shifting. CONCLUSIONS Shifting the frequency spectrum of speech has systematic effects that are in line with known properties of speech acoustics, but the ensuing difficulties cannot be predicted based on tonotopic mismatch alone. Difficulties are subject to substantial individual differences in the capacity to adjust acoustic-phonetic mapping. These results help to explain why speech recognition in CI listeners cannot be fully predicted by peripheral factors like electrode placement and spectral resolution; even among listeners with functionally equivalent auditory input, there is an additional factor of simply being able or unable to flexibly adjust acoustic-phonetic mapping. This individual variability could motivate precise treatment approaches guided by an individual's relative reliance on wideband frequency representation (even if it is mismatched) or limited frequency coverage whose tonotopy is preserved.
Collapse
|
10
|
Cucis PA, Berger-Vachon C, Thaï-Van H, Hermann R, Gallego S, Truy E. Word Recognition and Frequency Selectivity in Cochlear Implant Simulation: Effect of Channel Interaction. J Clin Med 2021; 10:jcm10040679. [PMID: 33578696 PMCID: PMC7916371 DOI: 10.3390/jcm10040679] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Revised: 02/02/2021] [Accepted: 02/05/2021] [Indexed: 11/16/2022] Open
Abstract
In cochlear implants (CI), spread of neural excitation may produce channel interaction. Channel interaction disturbs the spectral resolution and, among other factors, seems to impair speech recognition, especially in noise. In this study, two tests were performed with 20 adult normal-hearing (NH) subjects under different vocoded simulations. First, there was a measurement of word recognition in noise while varying the number of selected channels (4, 8, 12 or 16 maxima out of 20) and the degree of simulated channel interaction (“Low”, “Medium” and “High”). Then, there was an evaluation of spectral resolution function of the degree of simulated channel interaction, reflected by the sharpness (Q10dB) of psychophysical tuning curves (PTCs). The results showed a significant effect of the simulated channel interaction on word recognition but did not find an effect of the number of selected channels. The intelligibility decreased significantly for the highest degree of channel interaction. Similarly, the highest simulated channel interaction impaired significantly the Q10dB. Additionally, a strong intra-individual correlation between frequency selectivity and word recognition in noise was observed. Lastly, the individual changes in frequency selectivity were positively correlated with the changes in word recognition when the degree of interaction went from “Low” to “High”. To conclude, the degradation seen for the highest degree of channel interaction suggests a threshold effect on frequency selectivity and word recognition. The correlation between frequency selectivity and intelligibility in noise supports the hypothesis that PTCs Q10dB can account for word recognition in certain conditions. Moreover, the individual variations of performances observed among subjects suggest that channel interaction does not have the same effect on each individual. Finally, these results highlight the importance of taking into account subjects’ individuality and to evaluate channel interaction through the speech processor.
Collapse
Affiliation(s)
- Pierre-Antoine Cucis
- Integrative, Multisensory, Perception, Action and Cognition Team (IMPACT), Lyon Neuroscience Research Center, CRNL Inserm U1028, CNRS UMR5292, 69675 Bron, France; (R.H.); (E.T.)
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- ENT and Cervico-Facial Surgery Department, Edouard Herriot Hospital, Hospices Civils de Lyon, 69003 Lyon, France
- Correspondence: ; Tel.: +33-472-110-0518
| | - Christian Berger-Vachon
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- Brain Dynamics and Cognition Team (DYCOG), Lyon Neuroscience Research Center, CRNL Inserm U1028, CNRS UMR5292, 69675 Bron, France
- Biomechanics and Impact Mechanics Laboratory (LBMC), French Institute of Science and Technology for Transport, Development and Networks (IFSTTAR), Gustave Eiffel University, 69675 Bron, France
| | - Hung Thaï-Van
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- Paris Hearing Institute, Institut Pasteur, Inserm U1120, 75015 Paris, France
- Department of Audiology and Otoneurological Evaluation, Edouard Herriot Hospital, Hospices Civils de Lyon, 69003 Lyon, France
| | - Ruben Hermann
- Integrative, Multisensory, Perception, Action and Cognition Team (IMPACT), Lyon Neuroscience Research Center, CRNL Inserm U1028, CNRS UMR5292, 69675 Bron, France; (R.H.); (E.T.)
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- ENT and Cervico-Facial Surgery Department, Edouard Herriot Hospital, Hospices Civils de Lyon, 69003 Lyon, France
| | - Stéphane Gallego
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- Neuronal Dynamics and Audition Team (DNA), Laboratory of Cognitive Neuroscience (LNSC), CNRS UMR 7291, Aix-Marseille University, CEDEX 3, 13331 Marseille, France
| | - Eric Truy
- Integrative, Multisensory, Perception, Action and Cognition Team (IMPACT), Lyon Neuroscience Research Center, CRNL Inserm U1028, CNRS UMR5292, 69675 Bron, France; (R.H.); (E.T.)
- Claude Bernard Lyon 1 University, 69100 Villeurbanne, France; (C.B.-V.); (H.T.-V.); (S.G.)
- ENT and Cervico-Facial Surgery Department, Edouard Herriot Hospital, Hospices Civils de Lyon, 69003 Lyon, France
| |
Collapse
|
11
|
Konerding W, Arenberg JG, Kral A, Baumhoff P. Late electrically-evoked compound action potentials as markers for acute micro-lesions of spiral ganglion neurons. Hear Res 2020; 413:108057. [PMID: 32883545 DOI: 10.1016/j.heares.2020.108057] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/24/2020] [Revised: 07/13/2020] [Accepted: 08/11/2020] [Indexed: 12/20/2022]
Abstract
Cochlear implants (CIs) are the treatment of choice for profoundly hearing impaired people. It has been proposed that speech perception in CI users is influenced by the neural health (deafferentation, demyelination and degeneration) of the cochlea, which may be heterogeneous along an individual cochlea. Several options have been put forward to account for these local differences in neural health when fitting the speech processor settings, however with mixed results. The interpretation of the results is hampered by the fact that reliable markers of locally restricted changes in spiral ganglion neuron (SGN) health are lacking. The aim of the study was (i) to establish mechanical micro-lesions in the guinea pig as a model of heterogeneous SGN deafferentation and degeneration and (ii) to assess potential electrophysiological markers that can also be used in human subjects. First, we defined the extent of micro-lesions in normal hearing animals using acoustically-evoked compound action potentials (aCAPs); second, we measured electrically-evoked CAPs (eCAPs) before and after focal lesioning in neomycin-deafened and implanted animals. Therefore, we inserted guinea pig adjusted 6-contact CIs through a cochleostomy in the scala tympani. The eCAP was recorded from a ball electrode at the round window niche in response to monopolar or bipolar, 50 µs/phase biphasic pulses of alternating anodic- and cathodic-leading polarity. To exclude the large electrical artifact from the analysis, we focused on the late eCAP component. We systematically isolated the eCAP parameter that showed local pre- versus post-lesion changes and lesion-target specificity. Histological evaluation of the cleared cochleae revealed focal damage of an average size of 0.0036 mm3 with an apical-basal span of maximal 440 µm. We found that the threshold of the late N2P2 eCAP component was significantly elevated after lesioning when stimulating at basal (near the lesion), but not apical (distant to the lesion) CI contacts. To circumvent the potentially conflicting influence of the apical-basal gradient in eCAP thresholds, we used the polarity effect (PE=cathodic-anodic) as a relative measure. During monopolar stimulation, but not bipolar stimulation, the PE was sensitive to the lesion target and showed significantly better cathodic than anodic thresholds after soma lesions. We conclude that the difference in N2P2 thresholds in response to cathodic versus anodic-leading monopolar stimulation corresponds to the presence of SGN soma damage, and may therefore be a marker for SGN loss. We consider this electrophysiological estimate of local neural health a potentially relevant tool for human applications because of the temporal separation from the stimulation artifact and possible implementation into common eCAP measurements.
Collapse
Affiliation(s)
- Wiebke Konerding
- Department of Experimental Otology, Hannover Medical School, Stadtfelddamm 34, 30625 Hannover, Germany.
| | - Julie G Arenberg
- Massachusetts Eye and Ear Infirmary, Harvard Medical School, 25 Shattuck Street, Boston, MA 02115, USA.
| | - Andrej Kral
- Department of Experimental Otology, Hannover Medical School, Stadtfelddamm 34, 30625 Hannover, Germany; Cluster of Excellence "Hearing4all", Germany.
| | - Peter Baumhoff
- Department of Experimental Otology, Hannover Medical School, Stadtfelddamm 34, 30625 Hannover, Germany.
| |
Collapse
|
12
|
Abstract
OBJECTIVES Serial recall of digits is frequently used to measure short-term memory span in various listening conditions. However, the use of digits may mask the effect of low quality auditory input. Digits have high frequency and are phonologically distinct relative to one another, so they should be easy to identify even with low quality auditory input. In contrast, larger item sets reduce listener ability to strategically constrain their expectations, which should reduce identification accuracy and increase the time and/or cognitive resources needed for identification when auditory quality is low. This diminished accuracy and increased cognitive load should interfere with memory for sequences of items drawn from large sets. The goal of this work was to determine whether this predicted interaction between auditory quality and stimulus set in short-term memory exists, and if so, whether this interaction is associated with processing speed, vocabulary, or attention. DESIGN We compared immediate serial recall within young adults with normal hearing across unprocessed and vocoded listening conditions for multiple stimulus sets. Stimulus sets were lists of digits (1 to 9), consonant-vowel-consonant (CVC) words (chosen from a list of 60 words), and CVC nonwords (chosen from a list of 50 nonwords). Stimuli were unprocessed or vocoded with an eight-channel noise vocoder. To support interpretation of responses, words and nonwords were selected to minimize inclusion of multiple phonemes from within a confusion cluster. We also measured receptive vocabulary (Peabody Picture Vocabulary Test [PPVT-4]), sustained attention (test of variables of attention [TOVA]), and repetition speed for individual items from each stimulus set under both listening conditions. RESULTS Vocoding stimuli had no impact on serial recall of digits, but reduced memory span for words and nonwords. This reduction in memory span was attributed to an increase in phonological confusions for nonwords. However, memory span for vocoded word lists remained reduced even after accounting for common phonetic confusions, indicating that lexical status played an additional role across listening conditions. Principal components analysis found two components that explained 84% of the variance in memory span across conditions. Component one had similar load across all conditions, indicating that participants had an underlying memory capacity, which was common to all conditions. Component two was loaded by performance in the vocoded word and nonword conditions, representing the sensitivity of memory span to vocoding of these stimuli. The order in which participants completed listening conditions had a small effect on memory span that could not account for the effect of listening condition. Repetition speed was fastest for digits, slower for words, and slowest for nonwords. On average, vocoding slowed repetition speed for all stimuli, but repetition speed was not predictive of individual memory span. Vocabulary and attention showed no correlation with memory span. CONCLUSIONS Our results replicated previous findings that low quality auditory input can impair short-term memory, and demonstrated that this impairment is sensitive to stimulus set. Using multiple stimulus sets in degraded listening conditions can isolate memory capacity (in digit span) from impaired item identification (in word and nonword span), which may help characterize the relationship between memory and speech recognition in difficult listening conditions.
Collapse
Affiliation(s)
- Adam K. Bosen
- Boys Town National Research Hospital, Omaha, NE, USA
| | | |
Collapse
|
13
|
Bosen AK, Barry MF. Serial Recall Predicts Vocoded Sentence Recognition Across Spectral Resolutions. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:1282-1298. [PMID: 32213149 PMCID: PMC7242981 DOI: 10.1044/2020_jslhr-19-00319] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Purpose The goal of this study was to determine how various aspects of cognition predict speech recognition ability across different levels of speech vocoding within a single group of listeners. Method We tested the ability of young adults (N = 32) with normal hearing to recognize Perceptually Robust English Sentence Test Open-set (PRESTO) sentences that were degraded with a vocoder to produce different levels of spectral resolution (16, eight, and four carrier channels). Participants also completed tests of cognition (fluid intelligence, short-term memory, and attention), which were used as predictors of sentence recognition. Sentence recognition was compared across vocoder conditions, predictors were correlated with individual differences in sentence recognition, and the relationships between predictors were characterized. Results PRESTO sentence recognition performance declined with a decreasing number of vocoder channels, with no evident floor or ceiling performance in any condition. Individual ability to recognize PRESTO sentences was consistent relative to the group across vocoder conditions. Short-term memory, as measured with serial recall, was a moderate predictor of sentence recognition (ρ = 0.65). Serial recall performance was constant across vocoder conditions when measured with a digit span task. Fluid intelligence was marginally correlated with serial recall, but not sentence recognition. Attentional measures had no discernible relationship to sentence recognition and a marginal relationship with serial recall. Conclusions Verbal serial recall is a substantial predictor of vocoded sentence recognition, and this predictive relationship is independent of spectral resolution. In populations that show variable speech recognition outcomes, such as listeners with cochlear implants, it should be possible to account for the independent effects of spectral resolution and verbal serial recall in their speech recognition ability. Supplemental Material https://doi.org/10.23641/asha.12021051.
Collapse
|
14
|
Winn MB. Accommodation of gender-related phonetic differences by listeners with cochlear implants and in a variety of vocoder simulations. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:174. [PMID: 32006986 PMCID: PMC7341679 DOI: 10.1121/10.0000566] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2019] [Revised: 12/06/2019] [Accepted: 12/13/2019] [Indexed: 06/01/2023]
Abstract
Speech perception requires accommodation of a wide range of acoustic variability across talkers. A classic example is the perception of "sh" and "s" fricative sounds, which are categorized according to spectral details of the consonant itself, and also by the context of the voice producing it. Because women's and men's voices occupy different frequency ranges, a listener is required to make a corresponding adjustment of acoustic-phonetic category space for these phonemes when hearing different talkers. This pattern is commonplace in everyday speech communication, and yet might not be captured in accuracy scores for whole words, especially when word lists are spoken by a single talker. Phonetic accommodation for fricatives "s" and "sh" was measured in 20 cochlear implant (CI) users and in a variety of vocoder simulations, including those with noise carriers with and without peak picking, simulated spread of excitation, and pulsatile carriers. CI listeners showed strong phonetic accommodation as a group. Each vocoder produced phonetic accommodation except the 8-channel noise vocoder, despite its historically good match with CI users in word intelligibility. Phonetic accommodation is largely independent of linguistic factors and thus might offer information complementary to speech intelligibility tests which are partially affected by language processing.
Collapse
Affiliation(s)
- Matthew B Winn
- Department of Speech & Hearing Sciences, University of Minnesota, 164 Pillsbury Drive Southeast, Minneapolis, Minnesota 55455, USA
| |
Collapse
|
15
|
Reducing Simulated Channel Interaction Reveals Differences in Phoneme Identification Between Children and Adults With Normal Hearing. Ear Hear 2019; 40:295-311. [PMID: 29927780 DOI: 10.1097/aud.0000000000000615] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Channel interaction, the stimulation of overlapping populations of auditory neurons by distinct cochlear implant (CI) channels, likely limits the speech perception performance of CI users. This study examined the role of vocoder-simulated channel interaction in the ability of children with normal hearing (cNH) and adults with normal hearing (aNH) to recognize spectrally degraded speech. The primary aim was to determine the interaction between number of processing channels and degree of simulated channel interaction on phoneme identification performance as a function of age for cNH and to relate those findings to aNH and to CI users. DESIGN Medial vowel and consonant identification of cNH (age 8-17 years) and young aNH were assessed under six (for children) or nine (for adults) different conditions of spectral degradation. Stimuli were processed using a noise-band vocoder with 8, 12, and 15 channels and synthesis filter slopes of 15 (aNH only), 30, and 60 dB/octave (all NH subjects). Steeper filter slopes (larger numbers) simulated less electrical current spread and, therefore, less channel interaction. Spectrally degraded performance of the NH listeners was also compared with the unprocessed phoneme identification of school-aged children and adults with CIs. RESULTS Spectrally degraded phoneme identification improved as a function of age for cNH. For vowel recognition, cNH exhibited an interaction between the number of processing channels and vocoder filter slope, whereas aNH did not. Specifically, for cNH, increasing the number of processing channels only improved vowel identification in the steepest filter slope condition. Additionally, cNH were more sensitive to changes in filter slope. As the filter slopes increased, cNH continued to receive vowel identification benefit beyond where aNH performance plateaued or reached ceiling. For all NH participants, consonant identification improved with increasing filter slopes but was unaffected by the number of processing channels. Although cNH made more phoneme identification errors overall, their phoneme error patterns were similar to aNH. Furthermore, consonant identification of adults with CI was comparable to aNH listening to simulations with shallow filter slopes (15 dB/octave). Vowel identification of earlier-implanted pediatric ears was better than that of later-implanted ears and more comparable to cNH listening in conditions with steep filter slopes (60 dB/octave). CONCLUSIONS Recognition of spectrally degraded phonemes improved when simulated channel interaction was reduced, particularly for children. cNH showed an interaction between number of processing channels and filter slope for vowel identification. The differences observed between cNH and aNH suggest that identification of spectrally degraded phonemes continues to improve through adolescence and that children may benefit from reduced channel interaction beyond where adult performance has plateaued. Comparison to CI users suggests that early implantation may facilitate development of better phoneme discrimination.
Collapse
|
16
|
Rødvik AK, Tvete O, Torkildsen JVK, Wie OB, Skaug I, Silvola JT. Consonant and Vowel Confusions in Well-Performing Children and Adolescents With Cochlear Implants, Measured by a Nonsense Syllable Repetition Test. Front Psychol 2019; 10:1813. [PMID: 31474900 PMCID: PMC6702790 DOI: 10.3389/fpsyg.2019.01813] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2019] [Accepted: 07/22/2019] [Indexed: 12/31/2022] Open
Abstract
Although the majority of early implanted, profoundly deaf children with cochlear implants (CIs), will develop correct pronunciation if they receive adequate oral language stimulation, many of them have difficulties with perceiving minute details of speech. The main aim of this study is to measure the confusion of consonants and vowels in well-performing children and adolescents with CIs. The study also aims to investigate how age at onset of severe to profound deafness influences perception. The participants are 36 children and adolescents with CIs (18 girls), with a mean (SD) age of 11.6 (3.0) years (range: 5.9-16.0 years). Twenty-nine of them are prelingually deaf and seven are postlingually deaf. Two reference groups of normal-hearing (NH) 6- and 13-year-olds are included. Consonant and vowel perception is measured by repetition of 16 bisyllabic vowel-consonant-vowel nonsense words and nine monosyllabic consonant-vowel-consonant nonsense words in an open-set design. For the participants with CIs, consonants were mostly confused with consonants with the same voicing and manner, and the mean (SD) voiced consonant repetition score, 63.9 (10.6)%, was considerably lower than the mean (SD) unvoiced consonant score, 76.9 (9.3)%. There was a devoicing bias for the stops; unvoiced stops were confused with other unvoiced stops and not with voiced stops, and voiced stops were confused with both unvoiced stops and other voiced stops. The mean (SD) vowel repetition score was 85.2 (10.6)% and there was a bias in the confusions of [i:] and [y:]; [y:] was perceived as [i:] twice as often as [y:] was repeated correctly. Subgroup analyses showed no statistically significant differences between the consonant scores for pre- and postlingually deaf participants. For the NH participants, the consonant repetition scores were substantially higher and the difference between voiced and unvoiced consonant repetition scores considerably lower than for the participants with CIs. The participants with CIs obtained scores close to ceiling on vowels and real-word monosyllables, but their perception was substantially lower for voiced consonants. This may partly be related to limitations in the CI technology for the transmission of low-frequency sounds, such as insertion depth of the electrode and ability to convey temporal information.
Collapse
Affiliation(s)
- Arne Kirkhorn Rødvik
- Department of Special Needs Education, Institute of Educational Sciences, University of Oslo, Oslo, Norway.,Cochlear Implant Unit, Department of Otorhinolaryngology, Division of Surgery and Clinical Neuroscience, Oslo University Hospital, Oslo, Norway
| | - Ole Tvete
- Cochlear Implant Unit, Department of Otorhinolaryngology, Division of Surgery and Clinical Neuroscience, Oslo University Hospital, Oslo, Norway
| | - Janne von Koss Torkildsen
- Department of Special Needs Education, Institute of Educational Sciences, University of Oslo, Oslo, Norway
| | - Ona Bø Wie
- Department of Special Needs Education, Institute of Educational Sciences, University of Oslo, Oslo, Norway.,Cochlear Implant Unit, Department of Otorhinolaryngology, Division of Surgery and Clinical Neuroscience, Oslo University Hospital, Oslo, Norway
| | | | - Juha Tapio Silvola
- Department of Special Needs Education, Institute of Educational Sciences, University of Oslo, Oslo, Norway.,Cochlear Implant Unit, Department of Otorhinolaryngology, Division of Surgery and Clinical Neuroscience, Oslo University Hospital, Oslo, Norway.,Ear, Nose, and Throat Department, Division of Surgery, Akershus University Hospital, Lørenskog, Norway
| |
Collapse
|
17
|
DiNino M, Arenberg JG. Age-Related Performance on Vowel Identification and the Spectral-temporally Modulated Ripple Test in Children With Normal Hearing and With Cochlear Implants. Trends Hear 2019; 22:2331216518770959. [PMID: 29708065 PMCID: PMC5949928 DOI: 10.1177/2331216518770959] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Children’s performance on psychoacoustic tasks improves with age, but inadequate auditory input may delay this maturation. Cochlear implant (CI) users receive a degraded auditory signal with reduced frequency resolution compared with normal, acoustic hearing; thus, immature auditory abilities may contribute to the variation among pediatric CI users’ speech recognition scores. This study investigated relationships between age-related variables, spectral resolution, and vowel identification scores in prelingually deafened, early-implanted children with CIs compared with normal hearing (NH) children. All participants performed vowel identification and the Spectral-temporally Modulated Ripple Test (SMRT). Vowel stimuli for NH children were vocoded to simulate the reduced spectral resolution of CI hearing. Age positively predicted NH children’s vocoded vowel identification scores, but time with the CI was a stronger predictor of vowel recognition and SMRT performance of children with CIs. For both groups, SMRT thresholds were related to vowel identification performance, analogous to previous findings in adults. Sequential information analysis of vowel feature perception indicated greater transmission of duration-related information compared with formant features in both groups of children. In addition, the amount of F2 information transmitted predicted SMRT thresholds in children with NH and with CIs. Comparisons between the two CIs of bilaterally implanted children revealed disparate task performance levels and information transmission values within the same child. These findings indicate that adequate auditory experience contributes to auditory perceptual abilities of pediatric CI users. Further, factors related to individual CIs may be more relevant to psychoacoustic task performance than are the overall capabilities of the child.
Collapse
Affiliation(s)
- Mishaela DiNino
- 1 Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| | - Julie G Arenberg
- 1 Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| |
Collapse
|
18
|
Jahn KN, Arenberg JG. Polarity Sensitivity in Pediatric and Adult Cochlear Implant Listeners. Trends Hear 2019; 23:2331216519862987. [PMID: 31373266 PMCID: PMC6681263 DOI: 10.1177/2331216519862987] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2019] [Revised: 06/04/2019] [Accepted: 06/19/2019] [Indexed: 12/02/2022] Open
Abstract
Modeling data suggest that sensitivity to the polarity of an electrical stimulus may reflect the integrity of the peripheral processes of the spiral ganglion neurons. Specifically, better sensitivity to anodic (positive) current than to cathodic (negative) current could indicate peripheral process degeneration or demyelination. The goal of this study was to characterize polarity sensitivity in pediatric and adult cochlear implant listeners (41 ears). Relationships between polarity sensitivity at threshold and (a) polarity sensitivity at suprathreshold levels, (b) age-group, (c) preimplantation duration of deafness, and (d) phoneme perception were determined. Polarity sensitivity at threshold was defined as the difference in single-channel behavioral thresholds measured in response to each of two triphasic pulses, where the central high-amplitude phase was either cathodic or anodic. Lower thresholds in response to anodic than to cathodic pulses may suggest peripheral process degeneration. On the majority of electrodes tested, threshold and suprathreshold sensitivity was lower for anodic than for cathodic stimulation; however, dynamic range was often larger for cathodic than for anodic stimulation. Polarity sensitivity did not differ between child- and adult-implanted listeners. Adults with long preimplantation durations of deafness tended to have better sensitivity to anodic pulses on channels that were estimated to interface poorly with the auditory nerve; this was not observed in the child-implanted group. Across subjects, duration of deafness predicted phoneme perception performance. The results of this study suggest that subject- and electrode-dependent differences in polarity sensitivity may assist in developing customized cochlear implant programming interventions for child- and adult-implanted listeners.
Collapse
Affiliation(s)
- Kelly N. Jahn
- Department of Speech and Hearing
Sciences,
University
of Washington, Seattle, WA, USA
| | - Julie G. Arenberg
- Massachusetts Eye and Ear, Department of
Otolaryngology, Harvard Medical School, Boston, MA, USA
| |
Collapse
|
19
|
DeVries L, Arenberg JG. Current Focusing to Reduce Channel Interaction for Distant Electrodes in Cochlear Implant Programs. Trends Hear 2018; 22:2331216518813811. [PMID: 30488764 PMCID: PMC6277758 DOI: 10.1177/2331216518813811] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2018] [Revised: 10/17/2018] [Accepted: 10/26/2018] [Indexed: 11/17/2022] Open
Abstract
Speech understanding abilities are highly variable among cochlear implant (CI) listeners. Poor electrode-neuron interfaces (ENIs) caused by sparse neural survival or distant electrode placement may lead to increased channel interaction and reduced speech perception. Currently, it is not possible to directly measure neural survival in CI listeners; therefore, obtaining information about electrode position is an alternative approach to assessing ENIs. This information can be estimated with computerized tomography (CT) imaging; however, postoperative CT imaging is not often available. A reliable method to assess channel interaction, such as the psychophysical tuning curve (PTC), offers an alternative way to identify poor ENIs. This study aimed to determine (a) the within-subject relationship between CT-estimated electrode distance and PTC bandwidths, and (b) whether using focused stimulation on channels with suspected poor ENI improves vowel identification and sentence recognition. In 13 CI listeners, CT estimates of electrode-to-modiolus distance and PTCs bandwidths were measured for all available electrodes. Two test programs were created, wherein a subset of electrodes used focused stimulation based on (a) broad PTC bandwidth (Tuning) and (b) far electrode-to-modiolus distance (Distance). Two control programs were also created: (a) Those channels not focused in the Distance program (Inverse-Control), and (b) an all-channel monopolar program (Monopolar-Control). Across subjects, scores on the Distance and Tuning programs were significantly higher than the Inverse-Control program, and similar to the Monopolar-Control program. Subjective ratings were similar for all programs. These findings suggest that focusing channels suspected to have a high degree of channel interaction result in quite different outcomes, acutely.
Collapse
Affiliation(s)
- Lindsay DeVries
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Julie G. Arenberg
- Department of Speech and Hearing Sciences, University of Washington, Seattle, WA, USA
| |
Collapse
|
20
|
Zaar J, Schmitt N, Derleth RP, DiNino M, Arenberg JG, Dau T. Predicting effects of hearing-instrument signal processing on consonant perception. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:3216. [PMID: 29195458 DOI: 10.1121/1.5011737] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
This study investigated the influence of hearing-aid (HA) and cochlear-implant (CI) processing on consonant perception in normal-hearing (NH) listeners. Measured data were compared to predictions obtained with a speech perception model [Zaar and Dau (2017). J. Acoust. Soc. Am. 141, 1051-1064] that combines an auditory processing front end with a correlation-based template-matching back end. In terms of HA processing, effects of strong nonlinear frequency compression and impulse-noise suppression were measured in 10 NH listeners using consonant-vowel stimuli. Regarding CI processing, the consonant perception data from DiNino et al. [(2016). J. Acoust. Soc. Am. 140, 4404-4418] were considered, which were obtained with noise-vocoded vowel-consonant-vowel stimuli in 12 NH listeners. The inputs to the model were the same stimuli as were used in the corresponding experiments. The model predictions obtained for the two data sets showed a large agreement with the perceptual data both in terms of consonant recognition and confusions, demonstrating the model's sensitivity to supra-threshold effects of hearing-instrument signal processing on consonant perception. The results could be useful for the evaluation of hearing-instrument processing strategies, particularly when combined with simulations of individual hearing impairment.
Collapse
Affiliation(s)
- Johannes Zaar
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kongens Lyngby, Denmark
| | | | | | - Mishaela DiNino
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd Street, Box 354875, Seattle, Washington 98105, USA
| | - Julie G Arenberg
- Department of Speech and Hearing Sciences, University of Washington, 1417 NE 42nd Street, Box 354875, Seattle, Washington 98105, USA
| | - Torsten Dau
- Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, DK-2800 Kongens Lyngby, Denmark
| |
Collapse
|