1
|
Cychosz M, Winn MB, Goupell MJ. How to vocode: Using channel vocoders for cochlear-implant research. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2024; 155:2407-2437. [PMID: 38568143 PMCID: PMC10994674 DOI: 10.1121/10.0025274] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/19/2023] [Accepted: 02/23/2024] [Indexed: 04/05/2024]
Abstract
The channel vocoder has become a useful tool to understand the impact of specific forms of auditory degradation-particularly the spectral and temporal degradation that reflect cochlear-implant processing. Vocoders have many parameters that allow researchers to answer questions about cochlear-implant processing in ways that overcome some logistical complications of controlling for factors in individual cochlear implant users. However, there is such a large variety in the implementation of vocoders that the term "vocoder" is not specific enough to describe the signal processing used in these experiments. Misunderstanding vocoder parameters can result in experimental confounds or unexpected stimulus distortions. This paper highlights the signal processing parameters that should be specified when describing vocoder construction. The paper also provides guidance on how to determine vocoder parameters within perception experiments, given the experimenter's goals and research questions, to avoid common signal processing mistakes. Throughout, we will assume that experimenters are interested in vocoders with the specific goal of better understanding cochlear implants.
Collapse
Affiliation(s)
- Margaret Cychosz
- Department of Linguistics, University of California, Los Angeles, Los Angeles, California 90095, USA
| | - Matthew B Winn
- Department of Speech-Language-Hearing Sciences, University of Minnesota, Minneapolis, Minnesota 55455, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, College Park, Maryland 20742, USA
| |
Collapse
|
2
|
Nix EP, Thompson NJ, Brown KD, Dedmon MM, Selleck AM, Overton AB, Canfarotta MW, Dillon MT. Incidence of Cochlear Implant Electrode Contacts in the Functional Acoustic Hearing Region and the Influence on Speech Recognition with Electric-Acoustic Stimulation. Otol Neurotol 2023; 44:1004-1010. [PMID: 37758328 PMCID: PMC10840620 DOI: 10.1097/mao.0000000000004021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/03/2023]
Abstract
OBJECTIVES To investigate the incidence of electrode contacts within the functional acoustic hearing region in cochlear implant (CI) recipients and to assess its influence on speech recognition for electric-acoustic stimulation (EAS) users. STUDY DESIGN Retrospective review. SETTING Tertiary referral center. PATIENTS One hundred five CI recipients with functional acoustic hearing preservation (≤80 dB HL at 250 Hz). INTERVENTIONS Cochlear implantation with a 24-, 28-, or 31.5-mm lateral wall electrode array. MAIN OUTCOME MEASURES Angular insertion depth (AID) of individual contacts was determined from imaging. Unaided acoustic thresholds and AID were used to calculate the proximity of contacts to the functional acoustic hearing region. The association between proximity values and speech recognition in quiet and noise for EAS users at 6 months postactivation was reviewed. RESULTS Sixty percent of cases had one or more contacts within the functional acoustic hearing region. Proximity was not significantly associated with speech recognition in quiet. Better performance in noise was observed for cases with close correspondence between the most apical contact and the upper edge of residual hearing, with poorer results for increasing proximity values in either the basal or apical direction ( r14 = 0.48, p = 0.043; r18 = -0.41, p = 0.045, respectively). CONCLUSION There was a high incidence of electrode contacts within the functional acoustic hearing region, which is not accounted for with default mapping procedures. The variability in outcomes across EAS users with default maps may be due in part to electric-on-acoustic interference, electric frequency-to-place mismatch, and/or failure to stimulate regions intermediate between the most apical electrode contact and the functional acoustic hearing region.
Collapse
Affiliation(s)
- Evan P Nix
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | - Nicholas J Thompson
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | - Kevin D Brown
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | - Matthew M Dedmon
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | - A Morgan Selleck
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | | | - Michael W Canfarotta
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| | - Margaret T Dillon
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina School of Medicine, Chapel Hill, NC
| |
Collapse
|
3
|
Großmann W. Listening with an Ageing Brain - a Cognitive Challenge. Laryngorhinootologie 2023; 102:S12-S34. [PMID: 37130528 PMCID: PMC10184676 DOI: 10.1055/a-1973-3038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Hearing impairment has been recently identified as a major modifiable risk factor for cognitive decline in later life and has been becoming of increasing scientific interest. Sensory and cognitive decline are connected by complex bottom-up and top-down processes, a sharp distinction between sensation, perception, and cognition is impossible. This review provides a comprehensive overview on the effects of healthy and pathological aging on auditory as well as cognitive functioning on speech perception and comprehension, as well as specific auditory deficits in the 2 most common neurodegenerative diseases in old age: Alzheimer disease and Parkinson syndrome. Hypotheses linking hearing loss to cognitive decline are discussed, and current knowledge on the effect of hearing rehabilitation on cognitive functioning is presented. This article provides an overview of the complex relationship between hearing and cognition in old age.
Collapse
Affiliation(s)
- Wilma Großmann
- Universitätsmedizin Rostock, Klinik und Poliklinik für Hals-Nasen-Ohrenheilkunde,Kopf- und Halschirurgie "Otto Körner"
| |
Collapse
|
4
|
Schvartz-Leyzac KC, Giordani B, Pfingst BE. Association of Aging and Cognition With Complex Speech Understanding in Cochlear-Implanted Adults: Use of a Modified National Institutes of Health (NIH) Toolbox Cognitive Assessment. JAMA Otolaryngol Head Neck Surg 2023; 149:239-246. [PMID: 36701145 PMCID: PMC9880868 DOI: 10.1001/jamaoto.2022.4806] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2022] [Accepted: 12/01/2022] [Indexed: 01/27/2023]
Abstract
Importance The association between cognitive function and outcomes in cochlear implant (CI) users is not completely understood, partly because some cognitive tests are confounded by auditory status. It is important to determine appropriate cognitive tests to use in a cohort of CI recipients. Objective To provide proof-of-concept for using an adapted version of the National Institutes of Health (NIH) Toolbox Cognition Battery in a cohort of patients with CIs and to explore how hearing in noise with a CI is affected by cognitive status using the adapted test. Design, Setting, and Participants In this prognostic study, participants listened to sentences presented in a speech-shaped background noise. Cognitive tests consisted of 7 subtests of the NIH Toolbox Cognition Battery that were adapted for hearing impaired individuals by including written instructions and visual stimuli. Participants were prospectively recruited from and evaluated at a tertiary medical center. All participants had at least 6 months' experience with their CI. Main Outcomes and Measures The main outcomes were performance on the adapted cognitive test and a speech recognition in noise task. Results Participants were 20 adult perilingually or postlingually deafened CI users (50% male participants; median [range] age, 66 [26-80] years old). Performance on a sentence recognition in noise task was negatively associated with the chronological age of the listener (R2 = 0.29; β = 0.16; standard error, SE = 0.06; t = 2.63; 95% confidence interval, 0.03-0.27). Testing using the adapted version of the NIH Toolbox Cognition Battery revealed that a test of processing speed was also associated with performance, using a standardized score that accounted for contributions of other demographic factors (R2 = 0.28; 95% confidence interval, -0.42 to -0.05). Conclusions and Relevance In this prognostic study, older CI users showed poorer performance on a sentence-in-noise test compared with younger users. This poorer performance was correlated with a cognitive deficit in processing speed when cognitive function was assessed using a test battery adapted for participants with hearing loss. These results provide initial proof-of-concept results for using a standardized and adapted cognitive test battery in CI recipients.
Collapse
Affiliation(s)
- Kara C. Schvartz-Leyzac
- Kresge Hearing Research Institute, Department of Otolaryngology, University of Michigan Health Systems, Ann Arbor
- Hearing Rehabilitation Center, Department of Otolaryngology, University of Michigan Health Systems, Ann Arbor
- Medical University of South Carolina, Charleston
| | - Bruno Giordani
- Department of Psychiatry & Michigan Alzheimer’s Disease Center, University of Michigan Health Systems, Ann Arbor
| | - Bryan E. Pfingst
- Kresge Hearing Research Institute, Department of Otolaryngology, University of Michigan Health Systems, Ann Arbor
| |
Collapse
|
5
|
Moberly AC, Varadarajan VV, Tamati TN. Noise-Vocoded Sentence Recognition and the Use of Context in Older and Younger Adult Listeners. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2023; 66:365-381. [PMID: 36475738 PMCID: PMC10023188 DOI: 10.1044/2022_jslhr-22-00184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/26/2022] [Revised: 08/11/2022] [Accepted: 08/18/2022] [Indexed: 06/17/2023]
Abstract
PURPOSE When listening to speech under adverse conditions, older adults, even with "age-normal" hearing, face challenges that may lead to poorer speech recognition than their younger peers. Older listeners generally demonstrate poorer suprathreshold auditory processing along with aging-related declines in neurocognitive functioning that may impair their ability to compensate using "top-down" cognitive-linguistic functions. This study explored top-down processing in older and younger adult listeners, specifically the use of semantic context during noise-vocoded sentence recognition. METHOD Eighty-four adults with age-normal hearing (45 young normal-hearing [YNH] and 39 older normal-hearing [ONH] adults) participated. Participants were tested for recognition accuracy for two sets of noise-vocoded sentence materials: one that was semantically meaningful and the other that was syntactically appropriate but semantically anomalous. Participants were also tested for hearing ability and for neurocognitive functioning to assess working memory capacity, speed of lexical access, inhibitory control, and nonverbal fluid reasoning, as well as vocabulary knowledge. RESULTS The ONH and YNH listeners made use of semantic context to a similar extent. Nonverbal reasoning predicted recognition of both meaningful and anomalous sentences, whereas pure-tone average contributed additionally to anomalous sentence recognition. None of the hearing, neurocognitive, or language measures significantly predicted the amount of context gain, computed as the difference score between meaningful and anomalous sentence recognition. However, exploratory cluster analyses demonstrated four listener profiles and suggested that individuals may vary in the strategies used to recognize speech under adverse listening conditions. CONCLUSIONS Older and younger listeners made use of sentence context to similar degrees. Nonverbal reasoning was found to be a contributor to noise-vocoded sentence recognition. However, different listeners may approach the problem of recognizing meaningful speech under adverse conditions using different strategies based on their hearing, neurocognitive, and language profiles. These findings provide support for the complexity of bottom-up and top-down interactions during speech recognition under adverse listening conditions.
Collapse
Affiliation(s)
- Aaron C. Moberly
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
| | | | - Terrin N. Tamati
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands
| |
Collapse
|
6
|
Li MM, Moberly AC, Tamati TN. Factors affecting talker discrimination ability in adult cochlear implant users. JOURNAL OF COMMUNICATION DISORDERS 2022; 99:106255. [PMID: 35988314 PMCID: PMC10659049 DOI: 10.1016/j.jcomdis.2022.106255] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/13/2021] [Revised: 08/10/2022] [Accepted: 08/11/2022] [Indexed: 06/15/2023]
Abstract
INTRODUCTION Real-world speech communication involves interacting with many talkers with diverse voices and accents. Many adults with cochlear implants (CIs) demonstrate poor talker discrimination, which may contribute to real-world communication difficulties. However, the factors contributing to talker discrimination ability, and how discrimination ability relates to speech recognition outcomes in adult CI users are still unknown. The current study investigated talker discrimination ability in adult CI users, and the contributions of age, auditory sensitivity, and neurocognitive skills. In addition, the relation between talker discrimination ability and multiple-talker sentence recognition was explored. METHODS Fourteen post-lingually deaf adult CI users (3 female, 11 male) with ≥1 year of CI use completed a talker discrimination task. Participants listened to two monosyllabic English words, produced by the same talker or by two different talkers, and indicated if the words were produced by the same or different talkers. Nine female and nine male native English talkers were paired, resulting in same- and different-talker pairs as well as same-gender and mixed-gender pairs. Participants also completed measures of spectro-temporal processing, neurocognitive skills, and multiple-talker sentence recognition. RESULTS CI users showed poor same-gender talker discrimination, but relatively good mixed-gender talker discrimination. Older age and weaker neurocognitive skills, in particular inhibitory control, were associated with less accurate mixed-gender talker discrimination. Same-gender discrimination was significantly related to multiple-talker sentence recognition accuracy. CONCLUSION Adult CI users demonstrate overall poor talker discrimination ability. Individual differences in mixed-gender discrimination ability were related to age and neurocognitive skills, suggesting that these factors contribute to the ability to make use of available, degraded talker characteristics. Same-gender talker discrimination was associated with multiple-talker sentence recognition, suggesting that access to subtle talker-specific cues may be important for speech recognition in challenging listening conditions.
Collapse
Affiliation(s)
- Michael M Li
- The Ohio State University Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, Columbus, OH, USA
| | - Aaron C Moberly
- The Ohio State University Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, Columbus, OH, USA
| | - Terrin N Tamati
- The Ohio State University Wexner Medical Center, Department of Otolaryngology - Head & Neck Surgery, Columbus, OH, USA; University Medical Center Groningen, University of Groningen, Department of Otorhinolaryngology/Head and Neck Surgery, Groningen, the Netherlands.
| |
Collapse
|
7
|
Age-Related Changes in Voice Emotion Recognition by Postlingually Deafened Listeners With Cochlear Implants. Ear Hear 2022; 43:323-334. [PMID: 34406157 PMCID: PMC8847542 DOI: 10.1097/aud.0000000000001095] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
OBJECTIVES Identification of emotional prosody in speech declines with age in normally hearing (NH) adults. Cochlear implant (CI) users have deficits in the perception of prosody, but the effects of age on vocal emotion recognition by adult postlingually deaf CI users are not known. The objective of the present study was to examine age-related changes in CI users' and NH listeners' emotion recognition. DESIGN Participants included 18 CI users (29.6 to 74.5 years) and 43 NH adults (25.8 to 74.8 years). Participants listened to emotion-neutral sentences spoken by a male and female talker in five emotions (happy, sad, scared, angry, neutral). NH adults heard them in four conditions: unprocessed (full spectrum) speech, 16-channel, 8-channel, and 4-channel noise-band vocoded speech. The adult CI users only listened to unprocessed (full spectrum) speech. Sensitivity (d') to emotions and Reaction Times were obtained using a single-interval, five-alternative, forced-choice paradigm. RESULTS For NH participants, results indicated age-related declines in Accuracy and d', and age-related increases in Reaction Time in all conditions. Results indicated an overall deficit, as well as age-related declines in overall d' for CI users, but Reaction Times were elevated compared with NH listeners and did not show age-related changes. Analysis of Accuracy scores (hit rates) were generally consistent with d' data. CONCLUSIONS Both CI users and NH listeners showed age-related deficits in emotion identification. The CI users' overall deficit in emotion perception, and their slower response times, suggest impaired social communication which may in turn impact overall well-being, particularly so for older CI users, as lower vocal emotion recognition scores have been associated with poorer subjective quality of life in CI patients.
Collapse
|
8
|
Abstract
OBJECTIVES When auditory and visual speech information are presented together, listeners obtain an audiovisual (AV) benefit or a speech understanding improvement compared with auditory-only (AO) or visual-only (VO) presentations. Cochlear-implant (CI) listeners, who receive degraded speech input and therefore understand speech using primarily temporal information, seem to readily use visual cues and can achieve a larger AV benefit than normal-hearing (NH) listeners. It is unclear, however, if the AV benefit remains relatively large for CI listeners when trying to understand foreign-accented speech when compared with unaccented speech. Accented speech can introduce changes to temporal auditory cues and visual cues, which could decrease the usefulness of AV information. Furthermore, we sought to determine if the AV benefit was relatively larger in CI compared with NH listeners for both unaccented and accented speech. DESIGN AV benefit was investigated for unaccented and Spanish-accented speech by presenting English sentences in AO, VO, and AV conditions to 15 CI and 15 age- and performance-matched NH listeners. Performance matching between NH and CI listeners was achieved by varying the number of channels of a noise vocoder for the NH listeners. Because of the differences in age and hearing history of the CI listeners, the effects of listener-related variables on speech understanding performance and AV benefit were also examined. RESULTS AV benefit was observed for both unaccented and accented conditions and for both CI and NH listeners. The two groups showed similar performance for the AO and AV conditions, and the normalized AV benefit was relatively smaller for the accented than the unaccented conditions. In the CI listeners, older age was associated with significantly poorer performance with the accented speaker compared with the unaccented speaker. The negative impact of age was somewhat reduced by a significant improvement in performance with access to AV information. CONCLUSIONS When auditory speech information is degraded by CI sound processing, visual cues can be used to improve speech understanding, even in the presence of a Spanish accent. The AV benefit of the CI listeners closely matched that of the NH listeners presented with vocoded speech, which was unexpected given that CI listeners appear to rely more on visual information to communicate. This result is perhaps due to the one-to-one age and performance matching of the listeners. While aging decreased CI listener performance with the accented speaker, access to visual cues boosted performance and could partially overcome the age-related speech understanding deficits for the older CI listeners.
Collapse
|
9
|
Perception of Child-Directed Versus Adult-Directed Emotional Speech in Pediatric Cochlear Implant Users. Ear Hear 2021; 41:1372-1382. [PMID: 32149924 DOI: 10.1097/aud.0000000000000862] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Cochlear implants (CIs) are remarkable in allowing individuals with severe to profound hearing loss to perceive speech. Despite these gains in speech understanding, however, CI users often struggle to perceive elements such as vocal emotion and prosody, as CIs are unable to transmit the spectro-temporal detail needed to decode affective cues. This issue becomes particularly important for children with CIs, but little is known about their emotional development. In a previous study, pediatric CI users showed deficits in voice emotion recognition with child-directed stimuli featuring exaggerated prosody. However, the large intersubject variability and differential developmental trajectory known in this population incited us to question the extent to which exaggerated prosody would facilitate performance in this task. Thus, the authors revisited the question with both adult-directed and child-directed stimuli. DESIGN Vocal emotion recognition was measured using both child-directed (CDS) and adult-directed (ADS) speech conditions. Pediatric CI users, aged 7-19 years old, with no cognitive or visual impairments and who communicated through oral communication with English as the primary language participated in the experiment (n = 27). Stimuli comprised 12 sentences selected from the HINT database. The sentences were spoken by male and female talkers in a CDS or ADS manner, in each of the five target emotions (happy, sad, neutral, scared, and angry). The chosen sentences were semantically emotion-neutral. Percent correct emotion recognition scores were analyzed for each participant in each condition (CDS vs. ADS). Children also completed cognitive tests of nonverbal IQ and receptive vocabulary, while parents completed questionnaires of CI and hearing history. It was predicted that the reduced prosodic variations found in the ADS condition would result in lower vocal emotion recognition scores compared with the CDS condition. Moreover, it was hypothesized that cognitive factors, perceptual sensitivity to complex pitch changes, and elements of each child's hearing history may serve as predictors of performance on vocal emotion recognition. RESULTS Consistent with our hypothesis, pediatric CI users scored higher on CDS compared with ADS speech stimuli, suggesting that speaking with an exaggerated prosody-akin to "motherese"-may be a viable way to convey emotional content. Significant talker effects were also observed in that higher scores were found for the female talker for both conditions. Multiple regression analysis showed that nonverbal IQ was a significant predictor of CDS emotion recognition scores while Years using CI was a significant predictor of ADS scores. Confusion matrix analyses revealed a dependence of results on specific emotions; for the CDS condition's female talker, participants had high sensitivity (d' scores) to happy and low sensitivity to the neutral sentences while for the ADS condition, low sensitivity was found for the scared sentences. CONCLUSIONS In general, participants had higher vocal emotion recognition to the CDS condition which also had more variability in pitch and intensity and thus more exaggerated prosody, in comparison to the ADS condition. Results suggest that pediatric CI users struggle with vocal emotion perception in general, particularly to adult-directed speech. The authors believe these results have broad implications for understanding how CI users perceive emotions both from an auditory communication standpoint and a socio-developmental perspective.
Collapse
|
10
|
Abstract
OBJECTIVES Cochlear implants (CIs) are considered a safe and effective intervention for more severe degrees of hearing loss in adults of all ages. Although older CI users ≥65 years of age can obtain large benefits in speech understanding from a CI, there is a growing body of literature suggesting that older CI users may not perform as well as younger CI users. One reason for this potential age-related limitation could be that default CI stimulation settings are not optimal for older CI users. The goal of this study was to determine whether improvements in speech understanding were possible when CI users were programmed with nondefault stimulation rates and to determine whether lower-than-default stimulation rates improved older CI users' speech understanding. DESIGN Sentence recognition was measured acutely using different stimulation rates in 37 CI users ranging in age from 22 to 87 years. Maps were created using rates of 500, 720, 900, and 1200 pulses per second (pps) for each subject. An additional map using a rate higher than 1200 pps was also created for individuals who used a higher rate in their clinical processors. Thus, the clinical rate of each subject was also tested, including non-default rates above 1200 pps for Cochlear users and higher rates consistent with the manufacturer defaults for subjects implanted with Advanced Bionics and Med-El devices. Speech understanding performance was evaluated at each stimulation rate using AzBio and Perceptually Robust English Sentence Test Open-set (PRESTO) sentence materials tested in quiet and in noise. RESULTS For Cochlear-brand users, speech understanding performance using non-default rates was slightly poorer when compared with the default rate (900 pps). However, this effect was offset somewhat by age, in which older subjects were able to maintain comparable performance using a 500-pps map compared with the default rate map when listening to the more difficult PRESTO sentence material. Advanced Bionics and Med-El users showed modest improvements in their overall performance using 720 pps compared with the default rate (>1200 pps). On the individual-subject level, 10 subjects (11 ears) showed a significant effect of stimulation rate, with 8 of those ears performing best with a lower-than-default rate. CONCLUSIONS Results suggest that default stimulation rates are likely sufficient for many CI users, but some CI users at any age can benefit from a lower-than-default rate. Future work that provides experience with novel rates in everyday life has the potential to identify more individuals whose performance could be improved with changes to stimulation rate.
Collapse
|
11
|
Binaural Optimization of Cochlear Implants: Discarding Frequency Content Without Sacrificing Head-Shadow Benefit. Ear Hear 2021; 41:576-590. [PMID: 31436754 DOI: 10.1097/aud.0000000000000784] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Single-sided deafness cochlear-implant (SSD-CI) listeners and bilateral cochlear-implant (BI-CI) listeners gain near-normal levels of head-shadow benefit but limited binaural benefits. One possible reason for these limited binaural benefits is that cochlear places of stimulation tend to be mismatched between the ears. SSD-CI and BI-CI patients might benefit from a binaural fitting that reallocates frequencies to reduce interaural place mismatch. However, this approach could reduce monaural speech recognition and head-shadow benefit by excluding low- or high-frequency information from one ear. This study examined how much frequency information can be excluded from a CI signal in the poorer-hearing ear without reducing head-shadow benefits and how these outcomes are influenced by interaural asymmetry in monaural speech recognition. DESIGN Speech-recognition thresholds for sentences in speech-shaped noise were measured for 6 adult SSD-CI listeners, 12 BI-CI listeners, and 9 normal-hearing listeners presented with vocoder simulations. Stimuli were presented using nonindividualized in-the-ear or behind-the-ear head-related impulse-response simulations with speech presented from a 70° azimuth (poorer-hearing side) and noise from 70° (better-hearing side), thereby yielding a better signal-to-noise ratio (SNR) at the poorer-hearing ear. Head-shadow benefit was computed as the improvement in bilateral speech-recognition thresholds gained from enabling the CI in the poorer-hearing, better-SNR ear. High- or low-pass filtering was systematically applied to the head-related impulse-response-filtered stimuli presented to the poorer-hearing ear. For the SSD-CI listeners and SSD-vocoder simulations, only high-pass filtering was applied, because the CI frequency allocation would never need to be adjusted downward to frequency-match the ears. For the BI-CI listeners and BI-vocoder simulations, both low and high pass filtering were applied. The normal-hearing listeners were tested with two levels of performance to examine the effect of interaural asymmetry in monaural speech recognition (vocoder synthesis-filter slopes: 5 or 20 dB/octave). RESULTS Mean head-shadow benefit was smaller for the SSD-CI listeners (~7 dB) than for the BI-CI listeners (~14 dB). For SSD-CI listeners, frequencies <1236 Hz could be excluded; for BI-CI listeners, frequencies <886 or >3814 Hz could be excluded from the poorer-hearing ear without reducing head-shadow benefit. Bilateral performance showed greater immunity to filtering than monaural performance, with gradual changes in performance as a function of filter cutoff. Real and vocoder-simulated CI users with larger interaural asymmetry in monaural performance had less head-shadow benefit. CONCLUSIONS The "exclusion frequency" ranges that could be removed without diminishing head-shadow benefit are interpreted in terms of low importance in the speech intelligibility index and a small head-shadow magnitude at low frequencies. Although groups and individuals with greater performance asymmetry gained less head-shadow benefit, the magnitudes of these factors did not predict the exclusion frequency range. Overall, these data suggest that for many SSD-CI and BI-CI listeners, the frequency allocation for the poorer-ear CI can be shifted substantially without sacrificing head-shadow benefit, at least for energetic maskers. Considering the two ears together as a single system may allow greater flexibility in discarding redundant frequency content from a CI in one ear when considering bilateral programming solutions aimed at reducing interaural frequency mismatch.
Collapse
|
12
|
Jahn KN, DeVries L, Arenberg JG. Recovery from forward masking in cochlear implant listeners: Effects of age and the electrode-neuron interface. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:1633. [PMID: 33765782 PMCID: PMC8267874 DOI: 10.1121/10.0003623] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/17/2020] [Revised: 02/12/2021] [Accepted: 02/12/2021] [Indexed: 06/12/2023]
Abstract
Older adults exhibit deficits in auditory temporal processing relative to younger listeners. These age-related temporal processing difficulties may be further exacerbated in older adults with cochlear implant (CIs) when CI electrodes poorly interface with their target auditory neurons. The aim of this study was to evaluate the potential interaction between chronological age and the estimated quality of the electrode-neuron interface (ENI) on psychophysical forward masking recovery, a measure that reflects single-channel temporal processing abilities. Fourteen CI listeners (age 15 to 88 years) with Advanced Bionics devices participated. Forward masking recovery was assessed on two channels in each ear (i.e., the channels with the lowest and highest signal detection thresholds). Results indicated that the rate of forward masking recovery declined with advancing age, and that the effect of age was more pronounced on channels estimated to interface poorly with the auditory nerve. These findings indicate that the quality of the ENI can influence the time course of forward masking recovery for older CI listeners. Channel-to-channel variability in the ENI likely interacts with central temporal processing deficits secondary to auditory aging, warranting further study of programming and rehabilitative approaches tailored to older listeners.
Collapse
Affiliation(s)
- Kelly N Jahn
- Department of Speech and Hearing Sciences, University of Washington, Seattle, Washington 98105, USA
| | - Lindsay DeVries
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Julie G Arenberg
- Department of Speech and Hearing Sciences, University of Washington, Seattle, Washington 98105, USA
| |
Collapse
|
13
|
Goupell MJ, Draves GT, Litovsky RY. Recognition of vocoded words and sentences in quiet and multi-talker babble with children and adults. PLoS One 2020; 15:e0244632. [PMID: 33373427 PMCID: PMC7771688 DOI: 10.1371/journal.pone.0244632] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2020] [Accepted: 12/14/2020] [Indexed: 11/18/2022] Open
Abstract
A vocoder is used to simulate cochlear-implant sound processing in normal-hearing listeners. Typically, there is rapid improvement in vocoded speech recognition, but it is unclear if the improvement rate differs across age groups and speech materials. Children (8–10 years) and young adults (18–26 years) were trained and tested over 2 days (4 hours) on recognition of eight-channel noise-vocoded words and sentences, in quiet and in the presence of multi-talker babble at signal-to-noise ratios of 0, +5, and +10 dB. Children achieved poorer performance than adults in all conditions, for both word and sentence recognition. With training, vocoded speech recognition improvement rates were not significantly different between children and adults, suggesting that improvement in learning how to process speech cues degraded via vocoding is absent of developmental differences across these age groups and types of speech materials. Furthermore, this result confirms that the acutely measured age difference in vocoded speech recognition persists after extended training.
Collapse
Affiliation(s)
- Matthew J. Goupell
- Department of Hearing and Speech Sciences, University of Maryland, Maryland, MD, United States of America
- * E-mail:
| | - Garrison T. Draves
- Waisman Center, University of Wisconsin, Madison, WI, United States of America
| | - Ruth Y. Litovsky
- Waisman Center, University of Wisconsin, Madison, WI, United States of America
- Department of Communication Sciences and Disorders, University of Wisconsin, Madison, WI, United States of America
| |
Collapse
|
14
|
Xie Z, Gaskins CR, Shader MJ, Gordon-Salant S, Anderson S, Goupell MJ. Age-Related Temporal Processing Deficits in Word Segments in Adult Cochlear-Implant Users. Trends Hear 2020; 23:2331216519886688. [PMID: 31808373 PMCID: PMC6900735 DOI: 10.1177/2331216519886688] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022] Open
Abstract
Aging may limit speech understanding outcomes in cochlear-implant (CI) users.
Here, we examined age-related declines in auditory temporal processing as a
potential mechanism that underlies speech understanding deficits associated with
aging in CI users. Auditory temporal processing was assessed with a
categorization task for the words dish and ditch (i.e., identify each token as
the word dish or ditch) on a continuum of
speech tokens with varying silence duration (0 to 60 ms) prior to the final
fricative. In Experiments 1 and 2, younger CI (YCI), middle-aged CI (MCI), and
older CI (OCI) users participated in the categorization task across a range of
presentation levels (25 to 85 dB). Relative to YCI, OCI required longer silence
durations to identify ditch and exhibited reduced ability to distinguish the
words dish and ditch (shallower slopes in the categorization function).
Critically, we observed age-related performance differences only at higher
presentation levels. This contrasted with findings from normal-hearing listeners
in Experiment 3 that demonstrated age-related performance differences
independent of presentation level. In summary, aging in CI users appears to
degrade the ability to utilize brief temporal cues in word identification,
particularly at high levels. Age-specific CI programming may potentially improve
clinical outcomes for speech understanding performance by older CI
listeners.
Collapse
Affiliation(s)
- Zilong Xie
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Casey R Gaskins
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Maureen J Shader
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Sandra Gordon-Salant
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Samira Anderson
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, USA
| |
Collapse
|
15
|
Anderson S, Roque L, Gaskins CR, Gordon-Salant S, Goupell MJ. Age-Related Compensation Mechanism Revealed in the Cortical Representation of Degraded Speech. J Assoc Res Otolaryngol 2020; 21:373-391. [PMID: 32643075 DOI: 10.1007/s10162-020-00753-4] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2019] [Accepted: 04/05/2020] [Indexed: 02/08/2023] Open
Abstract
Older adults understand speech with comparative ease in quiet, but signal degradation can hinder speech understanding much more than it does in younger adults. This difficulty may result, in part, from temporal processing deficits related to the aging process and/or high-frequency hearing loss that can occur in listeners who have normal- or near-normal-hearing thresholds in the speech frequency range. Temporal processing deficits may manifest as degraded neural representation in peripheral and brainstem/midbrain structures that lead to compensation, or changes in response strength in auditory cortex. Little is understood about the process by which the neural representation of signals is improved or restored by age-related cortical compensation mechanisms. Therefore, we used vocoding to simulate spectral degradation to compare the behavioral and neural representation of words that contrast on a temporal dimension. Specifically, we used the closure duration of the silent interval between the vowel and the final affricate /t∫/ or fricative /ʃ/ of the words DITCH and DISH, respectively. We obtained perceptual identification functions and electrophysiological neural measures (frequency-following responses (FFR) and cortical auditory-evoked potentials (CAEPs)) to unprocessed and vocoded versions of these words in young normal-hearing (YNH), older normal- or near-normal-hearing (ONH), and older hearing-impaired (OHI) listeners. We found that vocoding significantly reduced the slope of the perceptual identification function in only the OHI listeners. In contrast to the limited effects of vocoding on perceptual performance, vocoding had robust effects on the FFRs across age groups, such that stimulus-to-response correlations and envelope magnitudes were significantly lower for vocoded vs. unprocessed conditions. Increases in the P1 peak amplitude for vocoded stimuli were found for both ONH and OHI listeners, but not for the YNH listeners. These results suggest that while vocoding substantially degrades early neural representation of speech stimuli in the midbrain, there may be cortical compensation in older listeners that is not seen in younger listeners.
Collapse
Affiliation(s)
- Samira Anderson
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, 20742, USA.
| | - Lindsey Roque
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, 20742, USA
| | - Casey R Gaskins
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, 20742, USA
| | - Sandra Gordon-Salant
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, 20742, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, MD, 20742, USA
| |
Collapse
|
16
|
Tejani VD, Brown CJ. Speech masking release in Hybrid cochlear implant users: Roles of spectral and temporal cues in electric-acoustic hearing. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:3667. [PMID: 32486815 PMCID: PMC7255813 DOI: 10.1121/10.0001304] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/16/2019] [Revised: 05/05/2020] [Accepted: 05/05/2020] [Indexed: 06/04/2023]
Abstract
When compared with cochlear implant (CI) users utilizing electric-only (E-Only) stimulation, CI users utilizing electric-acoustic stimulation (EAS) in the implanted ear show improved speech recognition in modulated noise relative to steady-state noise (i.e., speech masking release). It has been hypothesized, but not shown, that masking release is attributed to spectral resolution and temporal fine structure (TFS) provided by acoustic hearing. To address this question, speech masking release, spectral ripple density discrimination thresholds, and fundamental frequency difference limens (f0DLs) were evaluated in the acoustic-only (A-Only), E-Only, and EAS listening modes in EAS CI users. The spectral ripple and f0DL tasks are thought to reflect access to spectral and TFS cues, which could impact speech masking release. Performance in all three measures was poorest when EAS CI users were tested using the E-Only listening mode, with significant improvements in A-Only and EAS listening modes. f0DLs, but not spectral ripple density discrimination thresholds, significantly correlated with speech masking release when assessed in the EAS listening mode. Additionally, speech masking release correlated with AzBio sentence recognition in noise. The correlation between speech masking release and f0DLs likely indicates that TFS cues provided by residual hearing were used to obtain speech masking release, which aided sentence recognition in noise.
Collapse
Affiliation(s)
- Viral D Tejani
- Otolaryngology-Head and Neck Surgery, University of Iowa Hospitals and Clinics, 200 Hawkins Drive, 21003 Pomerantz Family Pavilion, Iowa City, Iowa 52242-1078, USA
| | - Carolyn J Brown
- Communication Sciences and Disorders, Wendell Johnson Speech and Hearing Center-127B, University of Iowa, 250 Hawkins Drive, Iowa City, Iowa 52242, USA
| |
Collapse
|
17
|
Abstract
OBJECTIVES Serial recall of digits is frequently used to measure short-term memory span in various listening conditions. However, the use of digits may mask the effect of low quality auditory input. Digits have high frequency and are phonologically distinct relative to one another, so they should be easy to identify even with low quality auditory input. In contrast, larger item sets reduce listener ability to strategically constrain their expectations, which should reduce identification accuracy and increase the time and/or cognitive resources needed for identification when auditory quality is low. This diminished accuracy and increased cognitive load should interfere with memory for sequences of items drawn from large sets. The goal of this work was to determine whether this predicted interaction between auditory quality and stimulus set in short-term memory exists, and if so, whether this interaction is associated with processing speed, vocabulary, or attention. DESIGN We compared immediate serial recall within young adults with normal hearing across unprocessed and vocoded listening conditions for multiple stimulus sets. Stimulus sets were lists of digits (1 to 9), consonant-vowel-consonant (CVC) words (chosen from a list of 60 words), and CVC nonwords (chosen from a list of 50 nonwords). Stimuli were unprocessed or vocoded with an eight-channel noise vocoder. To support interpretation of responses, words and nonwords were selected to minimize inclusion of multiple phonemes from within a confusion cluster. We also measured receptive vocabulary (Peabody Picture Vocabulary Test [PPVT-4]), sustained attention (test of variables of attention [TOVA]), and repetition speed for individual items from each stimulus set under both listening conditions. RESULTS Vocoding stimuli had no impact on serial recall of digits, but reduced memory span for words and nonwords. This reduction in memory span was attributed to an increase in phonological confusions for nonwords. However, memory span for vocoded word lists remained reduced even after accounting for common phonetic confusions, indicating that lexical status played an additional role across listening conditions. Principal components analysis found two components that explained 84% of the variance in memory span across conditions. Component one had similar load across all conditions, indicating that participants had an underlying memory capacity, which was common to all conditions. Component two was loaded by performance in the vocoded word and nonword conditions, representing the sensitivity of memory span to vocoding of these stimuli. The order in which participants completed listening conditions had a small effect on memory span that could not account for the effect of listening condition. Repetition speed was fastest for digits, slower for words, and slowest for nonwords. On average, vocoding slowed repetition speed for all stimuli, but repetition speed was not predictive of individual memory span. Vocabulary and attention showed no correlation with memory span. CONCLUSIONS Our results replicated previous findings that low quality auditory input can impair short-term memory, and demonstrated that this impairment is sensitive to stimulus set. Using multiple stimulus sets in degraded listening conditions can isolate memory capacity (in digit span) from impaired item identification (in word and nonword span), which may help characterize the relationship between memory and speech recognition in difficult listening conditions.
Collapse
|
18
|
Pals C, Sarampalis A, Beynon A, Stainsby T, Başkent D. Effect of Spectral Channels on Speech Recognition, Comprehension, and Listening Effort in Cochlear-Implant Users. Trends Hear 2020; 24:2331216520904617. [PMID: 32189585 PMCID: PMC7082863 DOI: 10.1177/2331216520904617] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
In favorable listening conditions, cochlear-implant (CI) users can reach high
speech recognition scores with as little as seven active electrodes. Here, we
hypothesized that even when speech recognition is high, additional spectral
channels may still benefit other aspects of speech perception, such as
comprehension and listening effort. Twenty-five adult, postlingually deafened CI
users, selected from two Dutch implant centers for high clinical word
identification scores, participated in two experiments. Experimental conditions
were created by varying the number of active electrodes of the CIs between 7 and
15. In Experiment 1, response times (RTs) on the secondary task in a dual-task
paradigm were used as an indirect measure of listening effort, and in Experiment
2, sentence verification task (SVT) accuracy and RTs were used to measure speech
comprehension and listening effort, respectively. Speech recognition was near
ceiling for all conditions tested, as intended by the design. However, the
dual-task paradigm failed to show the hypothesized decrease in RTs with
increasing spectral channels. The SVT did show a systematic improvement in both
speech comprehension and response speed across all conditions. In conclusion,
the SVT revealed additional benefits in both speech comprehension and listening
effort for conditions in which high speech recognition was already achieved.
Hence, adding spectral channels may provide benefits for CI listeners that may
not be reflected by traditional speech tests. The SVT is a relatively simple
task that is easy to implement and may therefore be a good candidate for
identifying such additional benefits in research or clinical settings.
Collapse
Affiliation(s)
- Carina Pals
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands.,Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, the Netherlands
| | | | - Andy Beynon
- Department of Otorhinolaryngology, Head and Neck Surgery, Hearing and Implants, Radboud University Medical Centre, Nijmegen, the Netherlands
| | | | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, the Netherlands.,Research School of Behavioral and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, the Netherlands
| |
Collapse
|
19
|
Spectral-Temporal Trade-Off in Vocoded Sentence Recognition: Effects of Age, Hearing Thresholds, and Working Memory. Ear Hear 2020; 41:1226-1235. [PMID: 32032222 DOI: 10.1097/aud.0000000000000840] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES Cochlear implant (CI) signal processing degrades the spectral components of speech. This requires CI users to rely primarily on temporal cues, specifically, amplitude modulations within the temporal envelope, to recognize speech. Auditory temporal processing ability for envelope modulations worsens with advancing age, which may put older CI users at a disadvantage compared with younger users. To evaluate how potential age-related limitations for processing temporal envelope modulations impact spectrally degraded sentence recognition, noise-vocoded sentences were presented to younger and older normal-hearing listeners in quiet. Envelope modulation rates were varied from 10 to 500 Hz by adjusting the low-pass filter cutoff frequency (LPF). The goal of this study was to evaluate if age impacts recognition of noise-vocoded speech and if this age-related limitation existed for a specific range of envelope modulation rates. DESIGN Noise-vocoded sentence recognition in quiet was measured as a function of number of spectral channels (4, 6, 8, and 12 channels) and LPF (10, 20, 50, 75, 150, 375, and 500 Hz) in 15 younger normal-hearing listeners and 15 older near-normal-hearing listeners. Hearing thresholds and working memory were assessed to determine the extent to which these factors were related to recognition of noise-vocoded sentences. RESULTS Younger listeners achieved significantly higher sentence recognition scores than older listeners overall. Performance improved in both groups as the number of spectral channels and LPF increased. As the number of spectral channels increased, the differences in sentence recognition scores between groups decreased. A spectral-temporal trade-off was observed in both groups in which performance in the 8- and 12-channel conditions plateaued with lower-frequency amplitude modulations compared with the 4- and 6-channel conditions. There was no interaction between age group and LPF, suggesting that both groups obtained similar improvements in performance with increasing LPF. The lack of an interaction between age and LPF may be due to the nature of the task of recognizing sentences in quiet. Audiometric thresholds were the only significant predictor of vocoded sentence recognition. Although performance on the working memory task declined with advancing age, working memory scores did not predict sentence recognition. CONCLUSIONS Younger listeners outperformed older listeners for recognizing noise-vocoded sentences in quiet. The negative impact of age was reduced when ample spectral information was available. Age-related limitations for recognizing vocoded sentences were not affected by the temporal envelope modulation rate of the signal, but instead, appear to be related to a generalized task limitation or to reduced audibility of the signal.
Collapse
|
20
|
Moberly AC, Reed J. Making Sense of Sentences: Top-Down Processing of Speech by Adult Cochlear Implant Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:2895-2905. [PMID: 31330118 PMCID: PMC6802905 DOI: 10.1044/2019_jslhr-h-18-0472] [Citation(s) in RCA: 35] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Revised: 03/20/2019] [Accepted: 04/12/2019] [Indexed: 05/03/2023]
Abstract
Purpose Speech recognition relies upon a listener's successful pairing of the acoustic-phonetic details from the bottom-up input with top-down linguistic processing of the incoming speech stream. When the speech is spectrally degraded, such as through a cochlear implant (CI), this role of top-down processing is poorly understood. This study explored the interactions of top-down processing, specifically the use of semantic context during sentence recognition, and the relative contributions of different neurocognitive functions during speech recognition in adult CI users. Method Data from 41 experienced adult CI users were collected and used in analyses. Participants were tested for recognition and immediate repetition of speech materials in the clear. They were asked to repeat 2 sets of sentence materials, 1 that was semantically meaningful and 1 that was syntactically appropriate but semantically anomalous. Participants also were tested on 4 visual measures of neurocognitive functioning to assess working memory capacity (Digit Span; Wechsler, 2004), speed of lexical access (Test of Word Reading Efficiency; Torgeson, Wagner, & Rashotte, 1999), inhibitory control (Stroop; Stroop, 1935), and nonverbal fluid reasoning (Raven's Progressive Matrices; Raven, 2000). Results Individual listeners' inhibitory control predicted recognition of meaningful sentences when controlling for performance on anomalous sentences, our proxy for the quality of the bottom-up input. Additionally, speed of lexical access and nonverbal reasoning predicted recognition of anomalous sentences. Conclusions Findings from this study identified inhibitory control as a potential mechanism at work when listeners make use of semantic context during sentence recognition. Moreover, speed of lexical access and nonverbal reasoning were associated with recognition of sentences that lacked semantic context. These results motivate the development of improved comprehensive rehabilitative approaches for adult patients with CIs to optimize use of top-down processing and underlying core neurocognitive functions.
Collapse
Affiliation(s)
- Aaron C. Moberly
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
| | - Jessa Reed
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
| |
Collapse
|
21
|
Colby S, Clayards M, Baum S. The Role of Lexical Status and Individual Differences for Perceptual Learning in Younger and Older Adults. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:1855-1874. [PMID: 30003232 DOI: 10.1044/2018_jslhr-s-17-0392] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/18/2017] [Accepted: 03/22/2018] [Indexed: 06/08/2023]
Abstract
PURPOSE This study examined whether older adults remain perceptually flexible when presented with ambiguities in speech in the absence of lexically disambiguating information. We expected older adults to show less perceptual learning when top-down information was not available. We also investigated whether individual differences in executive function predicted perceptual learning in older and younger adults. METHOD Younger (n = 31) and older adults (n = 27) completed 2 perceptual learning tasks composed of a pretest, exposure, and posttest phase. Both learning tasks exposed participants to clear and ambiguous speech tokens, but crucially, the lexically guided learning task provided disambiguating lexical information whereas the distributional learning task did not. Participants also performed several cognitive tasks to investigate individual differences in working memory, vocabulary, and attention-switching control. RESULTS We found that perceptual learning is maintained in older adults, but that learning may be stronger in contexts where top-down information is available. Receptive vocabulary scores predicted learning across both age groups and in both learning tasks. CONCLUSIONS Implicit learning is maintained with age across different learning conditions but remains stronger when lexically biasing information is available. We find that receptive vocabulary is relevant for learning in both types of learning tasks, suggesting the importance of vocabulary knowledge for adapting to ambiguities in speech.
Collapse
Affiliation(s)
- Sarah Colby
- School of Communication Sciences & Disorders, McGill University, Montréal, Québec, Canada
| | - Meghan Clayards
- School of Communication Sciences & Disorders, McGill University, Montréal, Québec, Canada
- Department of Linguistics, McGill University, Montréal, Québec, Canada
| | - Shari Baum
- School of Communication Sciences & Disorders, McGill University, Montréal, Québec, Canada
| |
Collapse
|
22
|
Effects of lifetime noise exposure on the middle-age human auditory brainstem response, tinnitus and speech-in-noise intelligibility. Hear Res 2018; 365:36-48. [DOI: 10.1016/j.heares.2018.06.003] [Citation(s) in RCA: 80] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/21/2017] [Revised: 05/25/2018] [Accepted: 06/08/2018] [Indexed: 01/03/2023]
|
23
|
Age-Related Differences in the Processing of Temporal Envelope and Spectral Cues in a Speech Segment. Ear Hear 2018; 38:e335-e342. [PMID: 28562426 DOI: 10.1097/aud.0000000000000447] [Citation(s) in RCA: 32] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES As people age, they experience reduced temporal processing abilities. This results in poorer ability to understand speech, particularly for degraded input signals. Cochlear implants (CIs) convey speech information via the temporal envelopes of a spectrally degraded input signal. Because there is an increasing number of older CI users, there is a need to understand how temporal processing changes with age. Therefore, the goal of this study was to quantify age-related reduction in temporal processing abilities when attempting to discriminate words based on temporal envelope information from spectrally degraded signals. DESIGN Younger normal-hearing (YNH) and older normal-hearing (ONH) participants were presented a continuum of speech tokens that varied in silence duration between phonemes (0 to 60 ms in 10-ms steps), and were asked to identify whether the stimulus was perceived more as the word "dish" or "ditch." Stimuli were vocoded using tonal carriers. The number of channels (1, 2, 4, 8, 16, and unprocessed) and temporal envelope low-pass filter cutoff frequency (50 and 400 Hz) were systematically varied. RESULTS For the unprocessed conditions, the YNH participants perceived the word ditch for smaller silence durations than the ONH participants, indicating that aging affects temporal processing abilities. There was no difference in performance between the unprocessed and 16-channel, 400-Hz vocoded stimuli. Decreasing the number of spectral channels caused decreased ability to distinguish dish and ditch. Decreasing the envelope cutoff frequency also caused decreased ability to distinguish dish and ditch. The overall pattern of results revealed that reductions in spectral and temporal information had a relatively larger effect on the ONH participants compared with the YNH participants. CONCLUSIONS Aging reduces the ability to utilize brief temporal cues in speech segments. Reducing spectral information-as occurs in a channel vocoder and in CI speech processing strategies-forces participants to use temporal envelope information; however, older participants are less capable of utilizing this information. These results suggest that providing as much spectral and temporal speech information as possible would benefit older CI users relatively more than younger CI users. In addition, the present findings help set expectations of clinical outcomes for speech understanding performance by adult CI users as a function of age.
Collapse
|
24
|
Aushana Y, Souffi S, Edeline JM, Lorenzi C, Huetz C. Robust Neuronal Discrimination in Primary Auditory Cortex Despite Degradations of Spectro-temporal Acoustic Details: Comparison Between Guinea Pigs with Normal Hearing and Mild Age-Related Hearing Loss. J Assoc Res Otolaryngol 2018; 19:163-180. [PMID: 29302822 PMCID: PMC5878150 DOI: 10.1007/s10162-017-0649-1] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2017] [Accepted: 12/11/2017] [Indexed: 01/04/2023] Open
Abstract
This study investigated to which extent the primary auditory cortex of young normal-hearing and mild hearing-impaired aged animals is able to maintain invariant representation of critical temporal-modulation features when sounds are submitted to degradations of fine spectro-temporal acoustic details. This was achieved by recording ensemble of cortical responses to conspecific vocalizations in guinea pigs with either normal hearing or mild age-related sensorineural hearing loss. The vocalizations were degraded using a tone vocoder. The neuronal responses and their discrimination capacities (estimated by mutual information) were analyzed at single recording and population levels. For normal-hearing animals, the neuronal responses decreased as a function of the number of the vocoder frequency bands, so did their discriminative capacities at the single recording level. However, small neuronal populations were found to be robust to the degradations induced by the vocoder. Similar robustness was obtained when broadband noise was added to exacerbate further the spectro-temporal distortions produced by the vocoder. A comparable pattern of robustness to degradations in fine spectro-temporal details was found for hearing-impaired animals. However, the latter showed an overall decrease in neuronal discrimination capacities between vocalizations in noisy conditions. Consistent with previous studies, these results demonstrate that the primary auditory cortex maintains robust neural representation of temporal envelope features for communication sounds under a large range of spectro-temporal degradations.
Collapse
Affiliation(s)
- Yonane Aushana
- Paris-Saclay Institute of Neurosciences (Neuro-PSI), CNRS UMR 9197, Orsay, France
- Université Paris-Sud, 91405 Orsay cedex, France
- Université Paris-Saclay, 91405 Orsay cedex, France
| | - Samira Souffi
- Paris-Saclay Institute of Neurosciences (Neuro-PSI), CNRS UMR 9197, Orsay, France
- Université Paris-Sud, 91405 Orsay cedex, France
- Université Paris-Saclay, 91405 Orsay cedex, France
| | - Jean-Marc Edeline
- Paris-Saclay Institute of Neurosciences (Neuro-PSI), CNRS UMR 9197, Orsay, France
- Université Paris-Sud, 91405 Orsay cedex, France
- Université Paris-Saclay, 91405 Orsay cedex, France
| | - Christian Lorenzi
- Laboratoire des Systèmes Perceptifs, UMR CNRS 8248, Département d’Etudes Cognitives, Ecole Normale Supérieure (ENS), Paris Sciences & Lettres Research University, 75005 Paris, France
| | - Chloé Huetz
- Paris-Saclay Institute of Neurosciences (Neuro-PSI), CNRS UMR 9197, Orsay, France
- Université Paris-Sud, 91405 Orsay cedex, France
- Université Paris-Saclay, 91405 Orsay cedex, France
| |
Collapse
|
25
|
Hay-McCutcheon MJ, Peterson NR, Pisoni DB, Kirk KI, Yang X, Parton J. Performance variability on perceptual discrimination tasks in profoundly deaf adults with cochlear implants. JOURNAL OF COMMUNICATION DISORDERS 2018; 72:122-135. [PMID: 29395103 PMCID: PMC5860985 DOI: 10.1016/j.jcomdis.2018.01.005] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/28/2017] [Revised: 01/09/2018] [Accepted: 01/22/2018] [Indexed: 06/07/2023]
Abstract
OBJECTIVES The purpose of this study was to evaluate performance on two challenging listening tasks, talker and regional accent discrimination, and to assess variables that could have affected the outcomes. STUDY DESIGN A prospective study using 35 adults with one cochlear implant (CI) or a CI and a contralateral hearing aid (bimodal hearing) was conducted. Adults completed talker and regional accent discrimination tasks. METHODS Two-alternative forced-choice tasks were used to assess talker and accent discrimination in a group of adults who ranged in age from 30 years old to 81 years old. RESULTS A large amount of performance variability was observed across listeners for both discrimination tasks. Three listeners successfully discriminated between talkers for both listening tasks, 14 participants successfully completed one discrimination task and 18 participants were not able to discriminate between talkers for either listening task. Some adults who used bimodal hearing benefitted from the addition of acoustic cues provided through a HA but for others the HA did not help with discrimination abilities. Acoustic speech feature analysis of the test signals indicated that both the talker speaking rate and the fundamental frequency (F0) helped with talker discrimination. For accent discrimination, findings suggested that access to more salient spectral cues was important for better discrimination performance. CONCLUSIONS The ability to perform challenging discrimination tasks successfully likely involves a number of complex interactions between auditory and non-auditory pre- and post-implant factors. To understand why some adults with CIs perform similarly to adults with normal hearing and others experience difficulty discriminating between talkers, further research will be required with larger populations of adults who use unilateral CIs, bilateral CIs and bimodal hearing.
Collapse
Affiliation(s)
- Marcia J Hay-McCutcheon
- The University of Alabama, Department of Communicative Disorders, Speech and Hearing Center, Box 870242, United States.
| | | | - David B Pisoni
- Indiana University, Department of Psychological and Brain Sciences, United States
| | - Karen Iler Kirk
- The University of Illinois at Urbana-Champaign, Department of Communication Sciences and Disorders, United States
| | - Xin Yang
- The University of Alabama, Information Systems, Statistics, Management Sciences, United States
| | - Jason Parton
- The University of Alabama, Information Systems, Statistics, Management Sciences, United States
| |
Collapse
|
26
|
Yumba WK. Cognitive Processing Speed, Working Memory, and the Intelligibility of Hearing Aid-Processed Speech in Persons with Hearing Impairment. Front Psychol 2017; 8:1308. [PMID: 28861009 PMCID: PMC5559705 DOI: 10.3389/fpsyg.2017.01308] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2016] [Accepted: 07/17/2017] [Indexed: 11/13/2022] Open
Abstract
Previous studies have demonstrated that successful listening with advanced signal processing in digital hearing aids is associated with individual cognitive capacity, particularly working memory capacity (WMC). This study aimed to examine the relationship between cognitive abilities (cognitive processing speed and WMC) and individual listeners’ responses to digital signal processing settings in adverse listening conditions. A total of 194 native Swedish speakers (83 women and 111 men), aged 33–80 years (mean = 60.75 years, SD = 8.89), with bilateral, symmetrical mild to moderate sensorineural hearing loss who had completed a lexical decision speed test (measuring cognitive processing speed) and semantic word-pair span test (SWPST, capturing WMC) participated in this study. The Hagerman test (capturing speech recognition in noise) was conducted using an experimental hearing aid with three digital signal processing settings: (1) linear amplification without noise reduction (NoP), (2) linear amplification with noise reduction (NR), and (3) non-linear amplification without NR (“fast-acting compression”). The results showed that cognitive processing speed was a better predictor of speech intelligibility in noise, regardless of the types of signal processing algorithms used. That is, there was a stronger association between cognitive processing speed and NR outcomes and fast-acting compression outcomes (in steady state noise). We observed a weaker relationship between working memory and NR, but WMC did not relate to fast-acting compression. WMC was a relatively weaker predictor of speech intelligibility in noise. These findings might have been different if the participants had been provided with training and or allowed to acclimatize to binary masking noise reduction or fast-acting compression.
Collapse
Affiliation(s)
- Wycliffe Kabaywe Yumba
- Department of Behavioral Sciences and Learning, Linköping UniversityLinköping, Sweden.,Linnaeus Centre HEAD, Swedish Institute for Disability Research, Linköping UniversityLinköping, Sweden
| |
Collapse
|
27
|
Mahajan Y, Kim J, Davis C. Older and younger adults' identification of sentences filtered with amplitude and frequency modulations in quiet and noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:EL190. [PMID: 28863573 DOI: 10.1121/1.4997603] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Adding frequency modulations (FM) cues to vocoded (AM) speech aids speech recognition for younger listeners. However, this may not be true for older listeners since they have poorer FM detection thresholds. We measured FM detection thresholds of young and older adults; and in a sentence context examined whether adding FM cues to vocoded speech would assist older adults. Young and old participants were presented vocoded sentences in quiet and multitalker-babble with/without FM cues. Older adults had elevated FM detection thresholds but received the same-size FM benefit as younger adults, showing that they have the capacity to benefit from FM speech cues.
Collapse
Affiliation(s)
- Yatin Mahajan
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia , ,
| | - Jeesun Kim
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia , ,
| | - Chris Davis
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, New South Wales, Australia , ,
| |
Collapse
|
28
|
Moberly AC, Houston DM, Harris MS, Adunka OF, Castellanos I. Verbal working memory and inhibition-concentration in adults with cochlear implants. Laryngoscope Investig Otolaryngol 2017; 2:254-261. [PMID: 29094068 PMCID: PMC5655567 DOI: 10.1002/lio2.90] [Citation(s) in RCA: 31] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2017] [Revised: 05/19/2017] [Accepted: 06/20/2017] [Indexed: 11/27/2022] Open
Abstract
Objectives Neurocognitive functions contribute to speech recognition in postlingual adults with cochlear implants (CIs). In particular, better verbal working memory (WM) on modality‐specific (auditory) WM tasks predicts better speech recognition. It remains unclear, however, whether this association can be attributed to basic underlying modality‐general neurocognitive functions, or whether it is solely a result of the degraded nature of auditory signals delivered by the CI. Three hypotheses were tested: 1) Both modality‐specific and modality‐general tasks of verbal WM would predict scores of sentence recognition in speech‐shaped noise; 2) Basic modality‐general neurocognitive functions of controlled fluency and inhibition‐concentration would predict both modality‐specific and modality‐general verbal WM; and 3) Scores on both tasks of verbal WM would mediate the effects of more basic neurocognitive functions on sentence recognition. Study Design Cross‐sectional study of 30 postlingual adults with CIs and thirty age‐matched normal‐hearing (NH) controls. Materials and Methods Participants were tested for sentence recognition in speech‐shaped noise, along with verbal WM using a modality‐general task (Reading Span) and an auditory modality‐specific task (Listening Span). Participants were also assessed for controlled fluency and inhibition‐concentration abilities. Results For CI users only, Listening Span scores predicted sentence recognition, and Listening Span scores mediated the effects of inhibition‐concentration on speech recognition. Scores on Reading Span were not related to sentence recognition for either group. Conclusion Inhibition‐concentration skills play an important role in CI users' sentence recognition skills, with effects mediated by modality‐specific verbal WM. Further studies will examine inhibition‐concentration and WM skills as novel targets for clinical intervention. Level of Evidence 4.
Collapse
Affiliation(s)
- Aaron C Moberly
- Department of Otolaryngology The Ohio State University Wexner Medical Center
| | - Derek M Houston
- Department of Otolaryngology The Ohio State University Wexner Medical Center
| | - Michael S Harris
- Department of Otolaryngology The Ohio State University Wexner Medical Center
| | - Oliver F Adunka
- Department of Otolaryngology The Ohio State University Wexner Medical Center
| | - Irina Castellanos
- Department of Otolaryngology The Ohio State University Wexner Medical Center
| |
Collapse
|
29
|
Rosemann S, Gießing C, Özyurt J, Carroll R, Puschmann S, Thiel CM. The Contribution of Cognitive Factors to Individual Differences in Understanding Noise-Vocoded Speech in Young and Older Adults. Front Hum Neurosci 2017. [PMID: 28638329 PMCID: PMC5461255 DOI: 10.3389/fnhum.2017.00294] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/26/2022] Open
Abstract
Noise-vocoded speech is commonly used to simulate the sensation after cochlear implantation as it consists of spectrally degraded speech. High individual variability exists in learning to understand both noise-vocoded speech and speech perceived through a cochlear implant (CI). This variability is partly ascribed to differing cognitive abilities like working memory, verbal skills or attention. Although clinically highly relevant, up to now, no consensus has been achieved about which cognitive factors exactly predict the intelligibility of speech in noise-vocoded situations in healthy subjects or in patients after cochlear implantation. We aimed to establish a test battery that can be used to predict speech understanding in patients prior to receiving a CI. Young and old healthy listeners completed a noise-vocoded speech test in addition to cognitive tests tapping on verbal memory, working memory, lexicon and retrieval skills as well as cognitive flexibility and attention. Partial-least-squares analysis revealed that six variables were important to significantly predict vocoded-speech performance. These were the ability to perceive visually degraded speech tested by the Text Reception Threshold, vocabulary size assessed with the Multiple Choice Word Test, working memory gauged with the Operation Span Test, verbal learning and recall of the Verbal Learning and Retention Test and task switching abilities tested by the Comprehensive Trail-Making Test. Thus, these cognitive abilities explain individual differences in noise-vocoded speech understanding and should be considered when aiming to predict hearing-aid outcome.
Collapse
Affiliation(s)
- Stephanie Rosemann
- Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany.,Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany
| | - Carsten Gießing
- Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany
| | - Jale Özyurt
- Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany
| | - Rebecca Carroll
- Cluster of Excellence "Hearing4all", Carl von Ossietzky Universität OldenburgOldenburg, Germany.,Institute of Dutch Studies, Carl von Ossietzky Universität OldenburgOldenburg, Germany
| | - Sebastian Puschmann
- Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany
| | - Christiane M Thiel
- Biological Psychology, Department of Psychology, European Medical School, Carl von Ossietzky Universität OldenburgOldenburg, Germany.,Cluster of Excellence "Hearing4all", Carl von Ossietzky Universität OldenburgOldenburg, Germany
| |
Collapse
|
30
|
Jaekel BN, Newman RS, Goupell MJ. Speech Rate Normalization and Phonemic Boundary Perception in Cochlear-Implant Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:1398-1416. [PMID: 28395319 PMCID: PMC5580678 DOI: 10.1044/2016_jslhr-h-15-0427] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/11/2015] [Revised: 05/04/2016] [Accepted: 10/14/2016] [Indexed: 05/29/2023]
Abstract
PURPOSE Normal-hearing (NH) listeners rate normalize, temporarily remapping phonemic category boundaries to account for a talker's speech rate. It is unknown if adults who use auditory prostheses called cochlear implants (CI) can rate normalize, as CIs transmit degraded speech signals to the auditory nerve. Ineffective adjustment to rate information could explain some of the variability in this population's speech perception outcomes. METHOD Phonemes with manipulated voice-onset-time (VOT) durations were embedded in sentences with different speech rates. Twenty-three CI and 29 NH participants performed a phoneme identification task. NH participants heard the same unprocessed stimuli as the CI participants or stimuli degraded by a sine vocoder, simulating aspects of CI processing. RESULTS CI participants showed larger rate normalization effects (6.6 ms) than the NH participants (3.7 ms) and had shallower (less reliable) category boundary slopes. NH participants showed similarly shallow slopes when presented acoustically degraded vocoded signals, but an equal or smaller rate effect in response to reductions in available spectral and temporal information. CONCLUSION CI participants can rate normalize, despite their degraded speech input, and show a larger rate effect compared to NH participants. CI participants may particularly rely on rate normalization to better maintain perceptual constancy of the speech signal.
Collapse
Affiliation(s)
- Brittany N. Jaekel
- Department of Hearing and Speech Sciences, University of Maryland, College Park
| | - Rochelle S. Newman
- Department of Hearing and Speech Sciences, University of Maryland, College Park
| | - Matthew J. Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park
| |
Collapse
|
31
|
Moberly AC, Harris MS, Boyce L, Nittrouer S. Speech Recognition in Adults With Cochlear Implants: The Effects of Working Memory, Phonological Sensitivity, and Aging. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:1046-1061. [PMID: 28384805 PMCID: PMC5548076 DOI: 10.1044/2016_jslhr-h-16-0119] [Citation(s) in RCA: 46] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/25/2016] [Revised: 08/30/2016] [Accepted: 10/14/2016] [Indexed: 05/12/2023]
Abstract
Purpose Models of speech recognition suggest that "top-down" linguistic and cognitive functions, such as use of phonotactic constraints and working memory, facilitate recognition under conditions of degradation, such as in noise. The question addressed in this study was what happens to these functions when a listener who has experienced years of hearing loss obtains a cochlear implant. Method Thirty adults with cochlear implants and 30 age-matched controls with age-normal hearing underwent testing of verbal working memory using digit span and serial recall of words. Phonological capacities were assessed using a lexical decision task and nonword repetition. Recognition of words in sentences in speech-shaped noise was measured. Results Implant users had only slightly poorer working memory accuracy than did controls and only on serial recall of words; however, phonological sensitivity was highly impaired. Working memory did not facilitate speech recognition in noise for either group. Phonological sensitivity predicted sentence recognition for implant users but not for listeners with normal hearing. Conclusion Clinical speech recognition outcomes for adult implant users relate to the ability of these users to process phonological information. Results suggest that phonological capacities may serve as potential clinical targets through rehabilitative training. Such novel interventions may be particularly helpful for older adult implant users.
Collapse
Affiliation(s)
- Aaron C. Moberly
- Department of Otolaryngology–Head and Neck Surgery, Wexner Medical Center, The Ohio State University, Columbus
| | - Michael S. Harris
- Department of Otolaryngology–Head and Neck Surgery, Wexner Medical Center, The Ohio State University, Columbus
| | - Lauren Boyce
- Department of Otolaryngology–Head and Neck Surgery, Wexner Medical Center, The Ohio State University, Columbus
| | - Susan Nittrouer
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville
| |
Collapse
|
32
|
Taitelbaum-Swead R, Fostick L. The Effect of Age and Type of Noise on Speech Perception under Conditions of Changing Context and Noise Levels. Folia Phoniatr Logop 2016; 68:16-21. [PMID: 27362521 DOI: 10.1159/000444749] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022] Open
Abstract
OBJECTIVE Everyday life includes fluctuating noise levels, resulting in continuously changing speech intelligibility. The study aims were: (1) to quantify the amount of decrease in age-related speech perception, as a result of increasing noise level, and (2) to test the effect of age on context usage at the word level (smaller amount of contextual cues). PATIENTS AND METHODS A total of 24 young adults (age 20-30 years) and 20 older adults (age 60-75 years) were tested. Meaningful and nonsense one-syllable consonant-vowel-consonant words were presented with the background noise types of speech noise (SpN), babble noise (BN), and white noise (WN), with a signal-to-noise ratio (SNR) of 0 and -5 dB. RESULTS Older adults had lower accuracy in SNR = 0, with WN being the most difficult condition for all participants. Measuring the change in speech perception when SNR decreased showed a reduction of 18.6-61.5% in intelligibility, with age effect only for BN. Both young and older adults used less phonemic context with WN, as compared to other conditions. CONCLUSION Older adults are more affected by an increasing noise level of fluctuating informational noise as compared to steady-state noise. They also use less contextual cues when perceiving monosyllabic words. Further studies should take into consideration that when presenting the stimulus differently (change in noise level, less contextual cues), other perceptual and cognitive processes are involved.
Collapse
|
33
|
Hay-McCutcheon MJ, Peterson NR, Rosado CA, Pisoni DB. Identification of acoustically similar and dissimilar vowels in profoundly deaf adults who use hearing aids and/or cochlear implants: some preliminary findings. Am J Audiol 2015; 23:57-70. [PMID: 23824440 DOI: 10.1044/1059-0889(2013/13-0009)] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022] Open
Abstract
PURPOSE In this study, the authors examined the effects of aging and residual hearing on the identification of acoustically similar and dissimilar vowels in adults with postlingual deafness who use hearing aids (HAs) and/or cochlear implants (CIs). METHOD The authors used two groups of acoustically similar and dissimilar vowels to assess vowel identification. Also, the Consonant-Nucleus-Consonant Word Recognition Test (Peterson & Lehiste, 1962) and sentences from the Hearing in Noise Test (Nilsson, Soli, & Sullivan, 1994) were administered. Forty CI recipients with postlingual deafness (ages 31-81 years) participated in the study. RESULTS Acoustically similar vowels were more difficult to identify than acoustically dissimilar vowels. With increasing age, performance deteriorated when identifying acoustically similar vowels. Vowel identification was also affected by the use of a contralateral HA and the degree of residual hearing prior to implantation. Moderate correlations were found between speech perception and vowel identification performance. CONCLUSIONS Identification performance was affected by the acoustic similarity of the vowels. Older adults experienced more difficulty identifying acoustically similar confusable vowels than did younger adults. The findings might lend support to the ease of language understanding model (Ronnberg, Rudner, Foo, & Lunner, 2008), which proposes that the quality and perceptual robustness of acoustic input affects speech perception.
Collapse
Affiliation(s)
| | | | | | - David B. Pisoni
- Indiana University School of Medicine, Indianapolis
- Indiana University Bloomington
| |
Collapse
|
34
|
Wöstmann M, Schröger E, Obleser J. Acoustic Detail Guides Attention Allocation in a Selective Listening Task. J Cogn Neurosci 2015; 27:988-1000. [DOI: 10.1162/jocn_a_00761] [Citation(s) in RCA: 29] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Abstract
The flexible allocation of attention enables us to perceive and behave successfully despite irrelevant distractors. How do acoustic challenges influence this allocation of attention, and to what extent is this ability preserved in normally aging listeners? Younger and healthy older participants performed a masked auditory number comparison while EEG was recorded. To vary selective attention demands, we manipulated perceptual separability of spoken digits from a masking talker by varying acoustic detail (temporal fine structure). Listening conditions were adjusted individually to equalize stimulus audibility as well as the overall level of performance across participants. Accuracy increased, and response times decreased with more acoustic detail. The decrease in response times with more acoustic detail was stronger in the group of older participants. The onset of the distracting speech masker triggered a prominent contingent negative variation (CNV) in the EEG. Notably, CNV magnitude decreased parametrically with increasing acoustic detail in both age groups. Within identical levels of acoustic detail, larger CNV magnitude was associated with improved accuracy. Across age groups, neuropsychological markers further linked early CNV magnitude directly to individual attentional capacity. Results demonstrate for the first time that, in a demanding listening task, instantaneous acoustic conditions guide the allocation of attention. Second, such basic neural mechanisms of preparatory attention allocation seem preserved in healthy aging, despite impending sensory decline.
Collapse
Affiliation(s)
- Malte Wöstmann
- 1Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
- 2International Max Planck Research School on Neuroscience of Communication, Leipzig, Germany
| | | | - Jonas Obleser
- 1Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
| |
Collapse
|
35
|
Neural alpha dynamics in younger and older listeners reflect acoustic challenges and predictive benefits. J Neurosci 2015; 35:1458-67. [PMID: 25632123 DOI: 10.1523/jneurosci.3250-14.2015] [Citation(s) in RCA: 97] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
Abstract
Speech comprehension in multitalker situations is a notorious real-life challenge, particularly for older listeners. Younger listeners exploit stimulus-inherent acoustic detail, but are they also actively predicting upcoming information? And further, how do older listeners deal with acoustic and predictive information? To understand the neural dynamics of listening difficulties and according listening strategies, we contrasted neural responses in the alpha-band (∼10 Hz) in younger (20-30 years, n = 18) and healthy older (60-70 years, n = 20) participants under changing task demands in a two-talker paradigm. Electroencephalograms were recorded while humans listened to two spoken digits against a distracting talker and decided whether the second digit was smaller or larger. Acoustic detail (temporal fine structure) and predictiveness (the degree to which the first digit predicted the second) varied orthogonally. Alpha power at widespread scalp sites decreased with increasing acoustic detail (during target digit presentation) but also with increasing predictiveness (in-between target digits). For older compared with younger listeners, acoustic detail had a stronger impact on task performance and alpha power modulation. This suggests that alpha dynamics plays an important role in the changes in listening behavior that occur with age. Last, alpha power variations resulting from stimulus manipulations (of acoustic detail and predictiveness) as well as task-independent overall alpha power were related to subjective listening effort. The present data show that alpha dynamics is a promising neural marker of individual difficulties as well as age-related changes in sensation, perception, and comprehension in complex communication situations.
Collapse
|
36
|
Shafiro V, Sheft S, Risley R, Gygi B. Effects of age and hearing loss on the intelligibility of interrupted speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2015; 137:745-56. [PMID: 25698009 PMCID: PMC4336257 DOI: 10.1121/1.4906275] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
How age and hearing loss affect the perception of interrupted speech may vary based on both the physical properties of preserved or obliterated speech fragments and individual listener characteristics. To investigate perceptual processes and interruption parameters influencing intelligibility across interruption rates, participants of different age and hearing status heard sentences interrupted by silence at either a single primary rate (0.5-8 Hz; 25%, 50%, 75% duty cycle) or at an additional concurrent secondary rate (24 Hz; 50% duty cycle). Although age and hearing loss significantly affected intelligibility, the ability to integrate sub-phonemic speech fragments produced by the fast secondary rate was similar in all listener groups. Age and hearing loss interacted with rate with smallest group differences observed at the lowest and highest interruption rates of 0.5 and 24 Hz. Furthermore, intelligibility of dual-rate gated sentences was higher than single-rate gated sentences with the same proportion of retained speech. Correlations of intelligibility of interrupted speech to pure-tone thresholds, age, or measures of working memory and auditory spectro-temporal pattern discrimination were generally low-to-moderate and mostly nonsignificant. These findings demonstrate rate-dependent effects of age and hearing loss on the perception of interrupted speech, suggesting complex interactions of perceptual processes across different time scales.
Collapse
Affiliation(s)
- Valeriy Shafiro
- Department of Communication Disorders and Sciences, Rush University Medical Center, 600 South Paulina Street, Suite 1012 AAC, Chicago, Illinois 60612
| | - Stanley Sheft
- Department of Communication Disorders and Sciences, Rush University Medical Center, 600 South Paulina Street, Suite 1012 AAC, Chicago, Illinois 60612
| | - Robert Risley
- Department of Communication Disorders and Sciences, Rush University Medical Center, 600 South Paulina Street, Suite 1012 AAC, Chicago, Illinois 60612
| | - Brian Gygi
- National Institute for Health Research, Nottingham Hearing Biomedical Research Unit, Nottingham, United Kingdom
| |
Collapse
|
37
|
Füllgrabe C, Moore BCJ, Stone MA. Age-group differences in speech identification despite matched audiometrically normal hearing: contributions from auditory temporal processing and cognition. Front Aging Neurosci 2015; 6:347. [PMID: 25628563 PMCID: PMC4292733 DOI: 10.3389/fnagi.2014.00347] [Citation(s) in RCA: 227] [Impact Index Per Article: 25.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2014] [Accepted: 12/23/2014] [Indexed: 11/13/2022] Open
Abstract
Hearing loss with increasing age adversely affects the ability to understand speech, an effect that results partly from reduced audibility. The aims of this study were to establish whether aging reduces speech intelligibility for listeners with normal audiograms, and, if so, to assess the relative contributions of auditory temporal and cognitive processing. Twenty-one older normal-hearing (ONH; 60-79 years) participants with bilateral audiometric thresholds ≤ 20 dB HL at 0.125-6 kHz were matched to nine young (YNH; 18-27 years) participants in terms of mean audiograms, years of education, and performance IQ. Measures included: (1) identification of consonants in quiet and in noise that was unmodulated or modulated at 5 or 80 Hz; (2) identification of sentences in quiet and in co-located or spatially separated two-talker babble; (3) detection of modulation of the temporal envelope (TE) at frequencies 5-180 Hz; (4) monaural and binaural sensitivity to temporal fine structure (TFS); (5) various cognitive tests. Speech identification was worse for ONH than YNH participants in all types of background. This deficit was not reflected in self-ratings of hearing ability. Modulation masking release (the improvement in speech identification obtained by amplitude modulating a noise background) and spatial masking release (the benefit obtained from spatially separating masker and target speech) were not affected by age. Sensitivity to TE and TFS was lower for ONH than YNH participants, and was correlated positively with speech-in-noise (SiN) identification. Many cognitive abilities were lower for ONH than YNH participants, and generally were correlated positively with SiN identification scores. The best predictors of the intelligibility of SiN were composite measures of cognition and TFS sensitivity. These results suggest that declines in speech perception in older persons are partly caused by cognitive and perceptual changes separate from age-related changes in audiometric sensitivity.
Collapse
Affiliation(s)
| | | | - Michael A. Stone
- School of Psychological Sciences, University of ManchesterManchester, UK
- Central Manchester NHS Hospitals Foundation TrustManchester, UK
| |
Collapse
|
38
|
Souza PE, Sirow L. Relating working memory to compression parameters in clinically fit hearing AIDS. Am J Audiol 2014; 23:394-401. [PMID: 25123548 PMCID: PMC4332516 DOI: 10.1044/2014_aja-14-0006] [Citation(s) in RCA: 36] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2014] [Accepted: 08/05/2014] [Indexed: 11/09/2022] Open
Abstract
PURPOSE Several laboratory studies have demonstrated that working memory may influence response to compression speed in controlled (i.e., laboratory) comparisons of compression. In this study, the authors explored whether the same relationship would occur under less controlled conditions, as might occur in a typical audiology clinic. METHOD Participants included 27 older adults who sought hearing care in a private practice audiology clinic. Working memory was measured for each participant using a reading span test. The authors examined the relationship between working memory and aided speech recognition in noise, using clinically fit hearing aids with a range of compression speeds. RESULTS Working memory, amount of hearing loss, and age each contributed to speech recognition, but the contribution depended on the speed of the compression processor. For fast-acting compression, the best performance was obtained by patients with high working memory. For slow-acting compression, speech recognition was affected by age and amount of hearing loss but was not affected by working memory. CONCLUSIONS Despite the expectation of greater variability from differences in compression implementation, number of compression channels, or attendant signal processing, the relationship between working memory and compression speed showed a similar pattern as results from more controlled, laboratory-based studies.
Collapse
|
39
|
Helfer KS, Freyman RL. Stimulus and listener factors affecting age-related changes in competing speech perception. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:748-759. [PMID: 25096109 PMCID: PMC4187459 DOI: 10.1121/1.4887463] [Citation(s) in RCA: 56] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/12/2013] [Revised: 06/19/2014] [Accepted: 06/24/2014] [Indexed: 05/29/2023]
Abstract
The purpose of this study was to examine associations among hearing thresholds, cognitive ability, and speech understanding in adverse listening conditions within and between groups of younger, middle-aged, and older adults. Participants repeated back sentences played in the presence of several types of maskers (syntactically similar and syntactically different competing speech from one or two other talkers, and steady-state speech-shaped noise). They also completed tests of auditory short-term/working memory, processing speed, and inhibitory ability. Results showed that group differences in accuracy of word identification and in error patterns differed depending upon the number of masking voices; specifically, older and middle-aged individuals had particular difficulty, relative to younger subjects, in the presence of a single competing message. However, the effect of syntactic similarity was consistent across subject groups. Hearing loss, short-term memory, processing speed, and inhibitory ability were each related to some aspects of performance by the middle-aged and older participants. Notably, substantial age-related changes in speech recognition were apparent within the group of middle-aged listeners.
Collapse
Affiliation(s)
- Karen S Helfer
- Department of Communication Disorders, University of Massachusetts Amherst, 358 North Pleasant Street, Amherst, Massachusetts 01003
| | - Richard L Freyman
- Department of Communication Disorders, University of Massachusetts Amherst, 358 North Pleasant Street, Amherst, Massachusetts 01003
| |
Collapse
|
40
|
Warner-Czyz AD, Houston DM, Hynan LS. Vowel discrimination by hearing infants as a function of number of spectral channels. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 135:3017-24. [PMID: 24815281 PMCID: PMC4109213 DOI: 10.1121/1.4870700] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/14/2012] [Revised: 03/26/2014] [Accepted: 03/27/2014] [Indexed: 05/15/2023]
Abstract
Reduced spectral resolution negatively impacts speech perception, particularly perception of vowels and consonant place. This study assessed impact of number of spectral channels on vowel discrimination by 6-month-old infants with normal hearing by comparing three listening conditions: Unprocessed speech, 32 channels, and 16 channels. Auditory stimuli (/ti/ and /ta/) were spectrally reduced using a noiseband vocoder and presented to infants with normal hearing via visual habituation. Results supported a significant effect of number of channels on vowel discrimination by 6-month-old infants. No differences emerged between unprocessed and 32-channel conditions in which infants looked longer during novel stimulus trials (i.e., discrimination). The 16-channel condition yielded a significantly different pattern: Infants demonstrated no significant difference in looking time to familiar vs novel stimulus trials, suggesting infants cannot discriminate /ti/ and /ta/ with only 16 channels. Results support effects of spectral resolution on vowel discrimination. Relative to published reports, young infants need more spectral detail than older children and adults to perceive spectrally degraded speech. Results have implications for development of perception by infants with hearing loss who receive auditory prostheses.
Collapse
Affiliation(s)
- Andrea D Warner-Czyz
- Department of Communication Sciences and Disorders, The University of Texas at Dallas, Callier Advanced Hearing Research Center, 1966 Inwood Road, Dallas, Texas 75235
| | - Derek M Houston
- Department of Otolaryngology, Head and Neck Surgery, Indiana University School of Medicine, 699 Riley Hospital Drive/RR044, Indianapolis, Indiana 46202
| | - Linda S Hynan
- Departments of Clinical Sciences and Psychiatry, The University of Texas Southwestern Medical Center, 5323 Harry Hines Boulevard, Dallas, Texas 75390
| |
Collapse
|
41
|
Ji C, Galvin JJ, Chang YP, Xu A, Fu QJ. Perception of speech produced by native and nonnative talkers by listeners with normal hearing and listeners with cochlear implants. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2014; 57:532-54. [PMID: 24686901 PMCID: PMC5847326 DOI: 10.1044/2014_jslhr-h-12-0404] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
PURPOSE The aim of this study was to evaluate the understanding of English sentences produced by native (English) and nonnative (Spanish) talkers by listeners with normal hearing (NH) and listeners with cochlear implants (CIs). METHOD Sentence recognition in noise was measured in adult subjects with CIs and subjects with NH, all of whom were native talkers of American English. Test sentences were from the Hearing in Noise Test (HINT) database and were produced in English by four native and eight nonnative talkers. Subjects also rated the intelligibility and accent for each talker. RESULTS The speech recognition thresholds in noise of subjects with CIs and subjects with NH were 4.23 dB and 1.32 dB poorer with nonnative talkers than with native talkers, respectively. Performance was significantly correlated with talker intelligibility and accent ratings for subjects with CIs but only correlated with talker intelligibility ratings for subjects with NH. For all subjects, performance with individual nonnative talkers was significantly correlated with talkers' number of years of residence in the United States. CONCLUSION CI users exhibited a larger deficit in speech understanding with nonnative talkers than did subjects with NH, relative to native talkers. Nonnative talkers' experience with native culture contributed strongly to speech understanding in noise, intelligibility ratings, and accent ratings of both listeners with NH and listeners with CIs.
Collapse
Affiliation(s)
- Caili Ji
- Department of Otolaryngology, Head and Neck Surgery, Qilu Hospital, Shandong University
- Communication and Neuroscience Division, House Ear Institute, 2100 W. 3 St., Los Angeles, CA 90057
| | - John J. Galvin
- Communication and Neuroscience Division, House Ear Institute, 2100 W. 3 St., Los Angeles, CA 90057
| | - Yi-ping Chang
- Communication and Neuroscience Division, House Ear Institute, 2100 W. 3 St., Los Angeles, CA 90057
| | - Anting Xu
- Department of Otolaryngology, Head and Neck Surgery, Qilu Hospital, Shandong University
| | - Qian-Jie Fu
- Communication and Neuroscience Division, House Ear Institute, 2100 W. 3 St., Los Angeles, CA 90057
| |
Collapse
|
42
|
Kates JM, Arehart KH, Souza PE. Integrating cognitive and peripheral factors in predicting hearing-aid processing effectiveness. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:4458. [PMID: 25669257 PMCID: PMC3874061 DOI: 10.1121/1.4824700] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/28/2013] [Revised: 09/16/2013] [Accepted: 09/26/2013] [Indexed: 05/28/2023]
Abstract
Individual factors beyond the audiogram, such as age and cognitive abilities, can influence speech intelligibility and speech quality judgments. This paper develops a neural network framework for combining multiple subject factors into a single model that predicts speech intelligibility and quality for a nonlinear hearing-aid processing strategy. The nonlinear processing approach used in the paper is frequency compression, which is intended to improve the audibility of high-frequency speech sounds by shifting them to lower frequency regions where listeners with high-frequency loss have better hearing thresholds. An ensemble averaging approach is used for the neural network to avoid the problems associated with overfitting. Models are developed for two subject groups, one having nearly normal hearing and the other mild-to-moderate sloping losses.
Collapse
Affiliation(s)
- James M Kates
- Department of Speech Language and Hearing Sciences, University of Colorado, Boulder, Colorado 80309
| | - Kathryn H Arehart
- Department of Speech Language and Hearing Sciences, University of Colorado, Boulder, Colorado 80309
| | - Pamela E Souza
- Department of Communication Sciences and Disorders and Knowles Hearing Center, Northwestern University, Evanston, Illinois 60201
| |
Collapse
|
43
|
Abstract
OBJECTIVES The purpose of this study was to determine the effect of measurement time of day on speech recognition in noise and the extent to which time-of-day effects differ with age. Older adults tend to have more difficulty understanding speech in noise than younger adults, even when hearing is normal. Two possible contributors to this age difference in speech recognition may be measurement time of day and inhibition. Most younger adults are "evening-type," showing peak circadian arousal in the evening, whereas most older adults are "morning-type," with circadian arousal peaking in the morning. Tasks that require inhibition of irrelevant information have been shown to be affected by measurement time of day, with maximum performance attained at one's peak time of day. The authors hypothesized that a change in inhibition will be associated with measurement time of day and therefore affect speech recognition in noise, with better performance in the morning for older adults and in the evening for younger adults. DESIGN Fifteen younger evening-type adults (20-28 years) and 15 older morning-type adults with normal hearing (66-78 years) listened to the Hearing in Noise Test (HINT) and the Quick Speech in Noise (QuickSIN) test in the morning and evening (peak and off-peak times). Time of day preference was assessed using the Morningness-Eveningness Questionnaire. Sentences and noise were presented binaurally through insert earphones. During morning and evening sessions, participants solved word-association problems within the visual-distraction task (VDT), which was used as an estimate of inhibition. After each session, participants rated perceived mental demand of the tasks using a revised version of the NASA Task Load Index. RESULTS Younger adults performed significantly better on the speech-in-noise tasks and rated themselves as requiring significantly less mental demand when tested at their peak (evening) than off-peak (morning) time of day. In contrast, time-of-day effects were not observed for the older adults on the speech recognition or rating tasks. Although older adults required significantly more advantageous signal-to-noise ratios than younger adults for equivalent speech-recognition performance, a significantly larger younger versus older age difference in speech recognition was observed in the evening than in the morning. Older adults performed significantly poorer than younger adults on the VDT, but performance was not affected by measurement time of day. VDT performance for misleading distracter items was significantly correlated with HINT and QuickSIN test performance at the peak measurement time of day. CONCLUSIONS Although all participants had normal hearing, speech recognition in noise was significantly poorer for older than younger adults, with larger age-related differences in the evening (an off-peak time for older adults) than in the morning. The significant effect of measurement time of day suggests that this factor may impact the clinical assessment of speech recognition in noise for all individuals. It appears that inhibition, as estimated by a visual distraction task for misleading visual items, is a cognitive mechanism that is related to speech-recognition performance in noise, at least at a listener's peak time of day.
Collapse
|
44
|
Abstract
OBJECTIVES Hearing aids use complex processing intended to improve speech recognition. Although many listeners benefit from such processing, it can also introduce distortion that offsets or cancels intended benefits for some individuals. The purpose of the present study was to determine the effects of cognitive ability (working memory) on individual listeners' responses to distortion caused by frequency compression applied to noisy speech. DESIGN The present study analyzed a large data set of intelligibility scores for frequency-compressed speech presented in quiet and at a range of signal-to-babble ratios. The intelligibility data set was based on scores from 26 adults with hearing loss with ages ranging from 62 to 92 years. The listeners were grouped based on working memory ability. The amount of signal modification (distortion) caused by frequency compression and noise was measured using a sound quality metric. Analysis of variance and hierarchical linear modeling were used to identify meaningful differences between subject groups as a function of signal distortion caused by frequency compression and noise. RESULTS Working memory was a significant factor in listeners' intelligibility of sentences presented in babble noise and processed with frequency compression based on sinusoidal modeling. At maximum signal modification (caused by both frequency compression and babble noise), the factor of working memory (when controlling for age and hearing loss) accounted for 29.3% of the variance in intelligibility scores. Combining working memory, age, and hearing loss accounted for a total of 47.5% of the variability in intelligibility scores. Furthermore, as the total amount of signal distortion increased, listeners with higher working memory performed better on the intelligibility task than listeners with lower working memory did. CONCLUSIONS Working memory is a significant factor in listeners' responses to total signal distortion caused by cumulative effects of babble noise and frequency compression implemented with sinusoidal modeling. These results, together with other studies focused on wide-dynamic range compression, suggest that older listeners with hearing loss and poor working memory are more susceptible to distortions caused by at least some types of hearing aid signal-processing algorithms and by noise, and that this increased susceptibility should be considered in the hearing aid fitting process.
Collapse
|
45
|
Zhang F, Benson C, Fu QJ. Cortical encoding of pitch contour changes in cochlear implant users: a mismatch negativity study. Audiol Neurootol 2013; 18:275-88. [PMID: 23920129 DOI: 10.1159/000351802] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2012] [Accepted: 04/26/2013] [Indexed: 11/19/2022] Open
Abstract
A better understanding of melodic pitch perception in cochlear implants (CIs) may guide signal processing and/or rehabilitation techniques to improve music perception and appreciation in CI patients. In this study, the mismatch negativity (MMN) in response to infrequent changes in 5-tone pitch contours was obtained in CI users and normal-hearing (NH) listeners. Melodic contour identification (MCI) was also measured. Results showed that MCI performance was poorer in CI than in NH subjects; the MMNs were missing in all CI subjects for the 1-semitone contours. The MMNs with the 5-semitone contours were observed in a smaller proportion of CI than NH subjects. Results suggest that encoding of pitch contour changes in CI users appears to be degraded, most likely due to the limited pitch cues provided by the CI and deafness-related compromise of brain substrates.
Collapse
Affiliation(s)
- Fawen Zhang
- Department of Communication Sciences and Disorders, University of Cincinnati, Cincinnati, Ohio, USA
| | | | | |
Collapse
|
46
|
McAuliffe MJ, Gibson EMR, Kerr SE, Anderson T, LaShell PJ. Vocabulary influences older and younger listeners' processing of dysarthric speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 134:1358-1368. [PMID: 23927132 DOI: 10.1121/1.4812764] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/02/2023]
Abstract
This study examined younger (n = 16) and older (n = 16) listeners' processing of dysarthric speech-a naturally occurring form of signal degradation. It aimed to determine how age, hearing acuity, memory, and vocabulary knowledge interacted in speech recognition and lexical segmentation. Listener transcripts were coded for accuracy and pattern of lexical boundary errors. For younger listeners, transcription accuracy was predicted by receptive vocabulary. For older listeners, this same effect existed but was moderated by pure-tone hearing thresholds. While both groups employed syllabic stress cues to inform lexical segmentation, older listeners were less reliant on this perceptual strategy. The results were interpreted to suggest that individuals with larger receptive vocabularies, with their presumed greater language familiarity, were better able to leverage cue redundancies within the speech signal to form lexical hypothesis-leading to an improved ability to comprehend dysarthric speech. This advantage was minimized as hearing thresholds increased. While the differing levels of reliance on stress cues across the listener groups could not be attributed to specific individual differences, it was hypothesized that some combination of larger vocabularies and reduced hearing thresholds in the older participant group led to them prioritize lexical cues as a segmentation frame.
Collapse
Affiliation(s)
- Megan J McAuliffe
- Department of Communication Disorders and New Zealand Institute of Language, Brain and Behaviour, University of Canterbury, Private Bag 4800, Christchurch 8140, New Zealand
| | | | | | | | | |
Collapse
|
47
|
Anderson S, White-Schwoch T, Parbery-Clark A, Kraus N. A dynamic auditory-cognitive system supports speech-in-noise perception in older adults. Hear Res 2013; 300:18-32. [PMID: 23541911 DOI: 10.1016/j.heares.2013.03.006] [Citation(s) in RCA: 154] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/30/2012] [Revised: 03/06/2013] [Accepted: 03/12/2013] [Indexed: 11/16/2022]
Abstract
Understanding speech in noise is one of the most complex activities encountered in everyday life, relying on peripheral hearing, central auditory processing, and cognition. These abilities decline with age, and so older adults are often frustrated by a reduced ability to communicate effectively in noisy environments. Many studies have examined these factors independently; in the last decade, however, the idea of an auditory-cognitive system has emerged, recognizing the need to consider the processing of complex sounds in the context of dynamic neural circuits. Here, we used structural equation modeling to evaluate the interacting contributions of peripheral hearing, central processing, cognitive ability, and life experiences to understanding speech in noise. We recruited 120 older adults (ages 55-79) and evaluated their peripheral hearing status, cognitive skills, and central processing. We also collected demographic measures of life experiences, such as physical activity, intellectual engagement, and musical training. In our model, central processing and cognitive function predicted a significant proportion of variance in the ability to understand speech in noise. To a lesser extent, life experience predicted hearing-in-noise ability through modulation of brainstem function. Peripheral hearing levels did not significantly contribute to the model. Previous musical experience modulated the relative contributions of cognitive ability and lifestyle factors to hearing in noise. Our models demonstrate the complex interactions required to hear in noise and the importance of targeting cognitive function, lifestyle, and central auditory processing in the management of individuals who are having difficulty hearing in noise.
Collapse
Affiliation(s)
- Samira Anderson
- Auditory Neuroscience Laboratory, Northwestern University, Evanston, IL 60208, USA
| | | | | | | |
Collapse
|
48
|
Aging and implantable hearing solutions. Abstracts from the Cochlear Science and Research Seminar. Paris, France. March 19-20, 2012. Audiol Neurootol 2012; 17 Suppl 1:1-26. [PMID: 22922653 DOI: 10.1159/000341356] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022] Open
|
49
|
Gender identification in younger and older adults: use of spectral and temporal cues in noise-vocoded speech. Ear Hear 2012; 33:411-20. [PMID: 22237163 DOI: 10.1097/aud.0b013e31823d78dc] [Citation(s) in RCA: 31] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
Abstract
OBJECTIVE The aim of this study was to investigate potential effects of age on the ability of normal-hearing (NH) adult listeners to utilize spectral and temporal cues when performing a voice gender identification task. DESIGN Ten younger and 10 older NH adult listeners were measured on their ability to correctly identify the speaker gender of six different vowel tokens (H-/vowel/-D) when spoken by eight speakers (four male and four female). Spectral (number of channels) and temporal cues (low-pass cut-off frequency for temporal envelope extraction) were systematically manipulated using noiseband vocoding techniques; stimuli contained 1, 4, 8, 16, or 32 spectral channels, while the low-pass cut-off frequency of the temporal envelope filter was 20, 50, 100, 200, or 400 Hz. Furthermore, the fundamental frequencies (F0s) of the vowel tokens were manipulated to create two conditions: "Expanded" (large range of F0 values) and "Compressed" (small range of F0 values). RESULTS In general, younger listeners performed better than the older listeners but only when stimuli were spectrally degraded. For both the Expanded and Compressed conditions, the overall performance of the younger listeners was better than that of the older listeners, suggesting age-related deficits in both spectral and temporal processing. Furthermore, a significant interaction between age group and temporal envelope cues revealed that older listeners received less benefit from increasing temporal envelope information compared with the benefit observed among younger listeners. In particular, the performance of the younger NH group (collapsed across number of channels), but not the older NH group, improved as the temporal envelope cut-off frequency was increased from 50 to 400 Hz. CONCLUSIONS The results reported here support previous findings of senescent declines in perceiving spectrally reduced speech and temporal amplitude modulation processing. These results suggest that when F0 values are similar to one another, younger listeners can use temporal cues alone to glean voice-pitch information but older listeners exhibit a lessened ability to use such cues. Previous studies have demonstrated the importance of temporal envelope cues in periodicity perception (e.g., gender recognition) by cochlear implant listeners. The results of this study suggest that aging affects the use of such cues, and consequently gender recognition might be poorer among older cochlear implant recipients.
Collapse
|
50
|
Shafiro V, Sheft S, Gygi B, Ho KTN. The influence of environmental sound training on the perception of spectrally degraded speech and environmental sounds. Trends Amplif 2012; 16:83-101. [PMID: 22891070 DOI: 10.1177/1084713812454225] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Perceptual training with spectrally degraded environmental sounds results in improved environmental sound identification, with benefits shown to extend to untrained speech perception as well. The present study extended those findings to examine longer-term training effects as well as effects of mere repeated exposure to sounds over time. Participants received two pretests (1 week apart) prior to a week-long environmental sound training regimen, which was followed by two posttest sessions, separated by another week without training. Spectrally degraded stimuli, processed with a four-channel vocoder, consisted of a 160-item environmental sound test, word and sentence tests, and a battery of basic auditory abilities and cognitive tests. Results indicated significant improvements in all speech and environmental sound scores between the initial pretest and the last posttest with performance increments following both exposure and training. For environmental sounds (the stimulus class that was trained), the magnitude of positive change that accompanied training was much greater than that due to exposure alone, with improvement for untrained sounds roughly comparable to the speech benefit from exposure. Additional tests of auditory and cognitive abilities showed that speech and environmental sound performance were differentially correlated with tests of spectral and temporal-fine-structure processing, whereas working memory and executive function were correlated with speech, but not environmental sound perception. These findings indicate generalizability of environmental sound training and provide a basis for implementing environmental sound training programs for cochlear implant (CI) patients.
Collapse
Affiliation(s)
- Valeriy Shafiro
- Department of Communication Disorders and Sciences, Rush University Medical Center, 600 S. Paulina Str., 1015 AAC, Chicago, IL 60612, USA.
| | | | | | | |
Collapse
|