1
|
Korhonen P, Kuk F, Slugocki C, Davis-Ruperto N. Use of Emotional and Neutral Speech in Evaluating Compression Speeds. J Am Acad Audiol 2021; 32:268-274. [PMID: 34034342 DOI: 10.1055/s-0041-1722945] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Academic Contribution Register] [Indexed: 10/21/2022]
Abstract
BACKGROUND Emotional speech differs from neutral speech in its envelope characteristics. Use of emotional speech materials may be more sensitive for evaluating signal processing algorithms that affect the temporal envelope. PURPOSE Subjective listener preference was compared between variable speed compression (VSC) and fast acting compression (FAC) amplitude compression algorithms using neutral and emotional speech. RESEARCH DESIGN The study used a single-blinded, repeated measures design. STUDY SAMPLE Twenty hearing-impaired (HI) listeners with a bilaterally symmetrical, mild- to-moderately severe sensorineural hearing loss and 21 listeners with normal hearing (NH) participated. INTERVENTION Speech was processed using FAC and VSC algorithms. DATA COLLECTION AND ANALYSIS A paired-comparison paradigm assessed subjective preference for FAC versus VSC using emotional and neutral speech materials. The significance of subjective preference for compression algorithm (FAC or VSC) was evaluated using a linear mixed effects model at each combination of stimulus type (emotional or neutral speech) and hearing group (NH or HI). RESULTS HI listeners showed a preference for VSC over FAC when listening to emotional speech. The same listeners showed a nonsignificant, preference for VSC over FAC when listening to neutral speech. NH listeners showed preference for VSC over FAC for both neutral and emotional speech materials. CONCLUSION These results suggest that the subjective sound quality of emotional speech is more susceptible than neutral speech to changes in the signal introduced by FAC. Clinicians should consider including emotional speech materials when evaluating listener preference for different compression speeds in the clinic.
Collapse
Affiliation(s)
- Petri Korhonen
- Office of Research in Clinical Amplification (ORCA-USA), WS Audiology, Lisle, Illinois
| | - Francis Kuk
- Office of Research in Clinical Amplification (ORCA-USA), WS Audiology, Lisle, Illinois
| | - Christopher Slugocki
- Office of Research in Clinical Amplification (ORCA-USA), WS Audiology, Lisle, Illinois
| | - Neal Davis-Ruperto
- Office of Research in Clinical Amplification (ORCA-USA), WS Audiology, Lisle, Illinois
| |
Collapse
|
2
|
Abstract
OBJECTIVES Previous research has shown that the association between hearing aid-processed speech recognition and individual working memory ability becomes stronger in more challenging conditions (e.g., higher background noise levels) and with stronger hearing aid processing (e.g., fast-acting wide dynamic range compression, WDRC). To date, studies have assumed omnidirectional microphone settings and collocated speech and noise conditions to study such relationships. Such conditions fail to recognize that most hearing aids are fit with directional processing that may improve the signal to noise ratio (SNR) and speech recognition in spatially separated speech and noise conditions. Here, we considered the possibility that directional processing may reduce the signal distortion arising from fast-acting WDRC and in turn influence the relationship between working memory ability and speech recognition with WDRC processing. The combined effects of hearing aid processing (WDRC and directionality) and SNR were quantified using a signal modification metric (cepstral correlation), which measures temporal envelope changes in the processed signal with respect to a linearly amplified reference. It was hypothesized that there will be a weaker association between working memory ability and speech recognition for hearing aid processing conditions that result in overall less signal modification (i.e., fewer changes to the processed envelope). DESIGN Twenty-three individuals with bilateral, mild to moderately severe sensorineural hearing loss participated in the study. Participants were fit with a commercially available hearing aid, and signal processing was varied in two dimensions: (1) Directionality (omnidirectional [OMNI] versus fixed-directional [DIR]), and (2) WDRC speed (fast-acting [FAST] versus slow-acting [SLOW]). Sentence recognition in spatially separated multi-talker babble was measured across a range of SNRs: 0 dB, 5 dB, 10 dB, and quiet. Cumulative signal modification was measured with individualized hearing aid settings, for all experimental conditions. A linear mixed-effects model was used to determine the relationship between speech recognition, working memory ability, and cumulative signal modification. RESULTS Signal modification results showed a complex relationship between directionality and WDRC speed, which varied by SNR. At 0 and 5 dB SNRs, signal modification was lower for SLOW than FAST regardless of directionality. However, at 10 dB SNR and in the DIR listening condition, there was no signal modification difference between FAST and SLOW. Consistent with previous studies, the association of speech recognition in noise with working memory ability depended on the level of signal modification. Contrary to the hypothesis above, however, there was a significant association of speech recognition with working memory only at lower levels of signal modification, and speech recognition increased at a faster rate for individuals with better working memory as signal modification decreased with DIR and SLOW. CONCLUSIONS This research suggests that working memory ability remains a significant predictor of speech recognition when WDRC and directionality are applied. Our findings revealed that directional processing can reduce the detrimental effect of fast-acting WDRC on speech cues at higher SNRs, which affects speech recognition ability. Contrary to some previous research, this study showed that individuals with better working memory ability benefitted more from a decrease in signal modification than individuals with poorer working memory ability.
Collapse
|
3
|
Kowalewski B, Dau T, May T. Perceptual Evaluation of Signal-to-Noise-Ratio-Aware Dynamic Range Compression in Hearing Aids. Trends Hear 2020; 24:2331216520930531. [PMID: 32573373 PMCID: PMC7313326 DOI: 10.1177/2331216520930531] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Academic Contribution Register] [Indexed: 11/15/2022] Open
Abstract
Dynamic range compression is a compensation strategy commonly used in modern hearing aids. Fast-acting systems respond relatively quickly to the fluctuations in the input level. This allows for more effective compression of the dynamic range of speech and hence enhanced the audibility of its low-intensity components. However, such processing also amplifies the background noise, distorts the modulation spectra of both the speech and the background, and can reduce the output signal-to-noise ratio (SNR). Recently, May et al. proposed a novel SNR-aware compression strategy, in which the compression speed is adapted depending on whether speech is present or absent. Fast-acting compression is applied to speech-dominated time–frequency (T-F) units, while noise-dominated T-F units are processed using slow-acting compression. It has been shown that this strategy provides a similar effective compression of the speech dynamic range as conventional fast-acting compression, while introducing fewer distortions of the modulation spectrum of the background and providing an improved output SNR. In this study, this SNR-aware compression strategy was compared with conventional fast- and slow-acting compression in terms of speech intelligibility and subjective preference in a group of 17 hearing-impaired listeners with varying degree of hearing loss. The results show a speech intelligibility benefit of the SNR-aware compression strategy over the conventional slow-acting system. Furthermore, the SNR-aware approach demonstrates an increased subjective preference compared with both conventional fast- and slow-acting systems.
Collapse
Affiliation(s)
- Borys Kowalewski
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark
| | - Tobias May
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark
| |
Collapse
|
4
|
Miller CW, Bernstein JGW, Zhang X, Wu YH, Bentler RA, Tremblay K. The Effects of Static and Moving Spectral Ripple Sensitivity on Unaided and Aided Speech Perception in Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:3113-3126. [PMID: 30515519 PMCID: PMC6440313 DOI: 10.1044/2018_jslhr-h-17-0373] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Academic Contribution Register] [Received: 10/02/2017] [Revised: 06/06/2018] [Accepted: 08/04/2018] [Indexed: 05/26/2023]
Abstract
PURPOSE This study evaluated whether certain spectral ripple conditions were more informative than others in predicting ecologically relevant unaided and aided speech outcomes. METHOD A quasi-experimental study design was used to evaluate 67 older adult hearing aid users with bilateral, symmetrical hearing loss. Speech perception in noise was tested under conditions of unaided and aided, auditory-only and auditory-visual, and 2 types of noise. Predictors included age, audiometric thresholds, audibility, hearing aid compression, and modulation depth detection thresholds for moving (4-Hz) or static (0-Hz) 2-cycle/octave spectral ripples applied to carriers of broadband noise or 2000-Hz low- or high-pass filtered noise. RESULTS A principal component analysis of the modulation detection data found that broadband and low-pass static and moving ripple detection thresholds loaded onto the first factor whereas high-pass static and moving ripple detection thresholds loaded onto a second factor. A linear mixed model revealed that audibility and the first factor (reflecting broadband and low-pass static and moving ripples) were significantly associated with speech perception performance. Similar results were found for unaided and aided speech scores. The interactions between speech conditions were not significant, suggesting that the relationship between ripples and speech perception was consistent regardless of visual cues or noise condition. High-pass ripple sensitivity was not correlated with speech understanding. CONCLUSIONS The results suggest that, for hearing aid users, poor speech understanding in noise and sensitivity to both static and slow-moving ripples may reflect deficits in the same underlying auditory processing mechanism. Significant factor loadings involving ripple stimuli with low-frequency content may suggest an impaired ability to use temporal fine structure information in the stimulus waveform. Support is provided for the use of spectral ripple testing to predict speech perception outcomes in clinical settings.
Collapse
Affiliation(s)
- Christi W. Miller
- Department of Speech and Hearing Sciences, University of Washington, Seattle
| | - Joshua G. W. Bernstein
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, MD
| | - Xuyang Zhang
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City
| | - Yu-Hsiang Wu
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City
| | - Ruth A. Bentler
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City
| | - Kelly Tremblay
- Department of Speech and Hearing Sciences, University of Washington, Seattle
| |
Collapse
|
5
|
Alexander JM, Rallapalli V. Acoustic and perceptual effects of amplitude and frequency compression on high-frequency speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:908. [PMID: 28863610 DOI: 10.1121/1.4997938] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Academic Contribution Register] [Indexed: 06/07/2023]
Abstract
This study investigated how six different amplification methods influence acoustic properties, and subsequently perception, of high-frequency cues in fricatives that have been processed with conventional full bandwidth amplification or nonlinear frequency compression (NFC)-12 conditions total. Amplification methods included linear gain, fast/slow-acting wide dynamic range compression crossed with fixed/individualized compression parameters, and a method with adaptive time constants. Twenty-one hearing-impaired listeners identified seven fricatives in nonsense syllables produced by female talkers. For NFC stimuli, frequency-compressed filters that precisely aligned 1/3-octave bands between input and output were used to quantify effective compression ratio, audibility, and temporal envelope modulation relative to the input. Results indicated significant relationships between these acoustic properties, each of which contributed significantly to fricative recognition across the entire corpus of stimuli. Recognition was significantly better for NFC stimuli compared with full bandwidth stimuli, regardless of the amplification method, which had complementary effects on audibility and envelope modulation. Furthermore, while there were significant differences in recognition across the amplification methods, they were not consistent across phonemes. Therefore, neither recognition nor acoustic data overwhelmingly suggest that one amplification method should be used over another for transmission of high-frequency cues in isolated syllables. Longer duration stimuli and more realistic listening conditions should be examined.
Collapse
Affiliation(s)
- Joshua M Alexander
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, Indiana 47907, USA
| | - Varsha Rallapalli
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, Indiana 47907, USA
| |
Collapse
|
6
|
Reinhart PN, Souza PE. Intelligibility and Clarity of Reverberant Speech: Effects of Wide Dynamic Range Compression Release Time and Working Memory. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2016; 59:1543-1554. [PMID: 27997667 PMCID: PMC5399768 DOI: 10.1044/2016_jslhr-h-15-0371] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Academic Contribution Register] [Received: 10/28/2015] [Revised: 01/24/2016] [Accepted: 05/23/2016] [Indexed: 05/31/2023]
Abstract
PURPOSE The purpose of this study was to examine the effects of varying wide dynamic range compression (WDRC) release time on intelligibility and clarity of reverberant speech. The study also considered the role of individual working memory. METHOD Thirty older listeners with mild to moderately-severe sloping sensorineural hearing loss participated. Individuals were divided into high and low working memory groups on the basis of the results of a reading span test. Participants listened binaurally to sentence stimuli simulated at a range of reverberation conditions and WDRC release times using a high compression ratio. Outcome measures included objective intelligibility and subjective clarity ratings. RESULTS Speech intelligibility and clarity ratings both decreased as a function of reverberation. The low working memory group demonstrated a greater decrease in intelligibility with increasing amounts of reverberation than the high working memory group. Both groups, regardless of working memory, had higher speech intelligibility and clarity ratings with longer WDRC release times. WDRC release time had a larger effect on speech intelligibility under more reverberant conditions. CONCLUSIONS Reverberation significantly affects speech intelligibility, particularly for individuals with lower working memory. In addition, longer release times in hearing aids may improve listener speech intelligibility and clarity in reverberant environments.
Collapse
Affiliation(s)
- Paul N. Reinhart
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL
| | - Pamela E. Souza
- Department of Communication Sciences and Disorders, Northwestern University, Evanston, IL
- Knowles Hearing Center, Evanston, IL
| |
Collapse
|
7
|
Abstract
OBJECTIVES The purpose of this study was to investigate the joint effects that wide dynamic range compression (WDRC) release time (RT) and number of channels have on recognition of sentences in the presence of steady and modulated maskers at different signal-to-noise ratios (SNRs). How the different combinations of WDRC parameters affect output SNR and the role this plays in the observed findings were also investigated. DESIGN Twenty-four listeners with mild to moderate sensorineural hearing loss identified sentences mixed with steady or modulated maskers at three SNRs (-5, 0, and +5 dB) that had been processed using a hearing aid simulator with six combinations of RT (40 and 640 msec) and number of channels (4, 8, and 16). Compression parameters were set using the Desired Sensation Level v5.0a prescriptive fitting method. For each condition, amplified speech and masker levels and the resultant long-term output SNR were measured. RESULTS Speech recognition with WDRC depended on the combination of RT and number of channels, with the greatest effects observed at 0 dB input SNR, in which mean speech recognition scores varied by 10 to 12% across WDRC manipulations. Overall, effect sizes were generally small. Across both masker types and the three SNRs tested, the best speech recognition was obtained with eight channels, regardless of RT. Increased speech levels, which favor audibility, were associated with the short RT and with an increase in the number of channels. These same conditions also increased masker levels by an even greater amount, for a net decrease in the long-term output SNR. Changes in long-term SNR across WDRC conditions were found to be strongly associated with changes in the temporal envelope shape as quantified by the Envelope Difference Index; however, neither of these factors fully explained the observed differences in speech recognition. CONCLUSIONS A primary finding of this study was that the number of channels had a modest effect when analyzed at each level of RT, with results suggesting that selecting eight channels for a given RT might be the safest choice. Effects were smaller for RT, with results suggesting that short RT was slightly better when only 4 channels were used and that long RT was better when 16 channels were used. Individual differences in how listeners were influenced by audibility, output SNR, temporal distortion, and spectral distortion may have contributed to the size of the effects found in this study. Because only general suppositions could made for how each of these factors may have influenced the overall results of this study, future research would benefit from exploring the predictive value of these and other factors in selecting the processing parameters that maximize speech recognition for individuals.
Collapse
|
8
|
How neuroscience relates to hearing aid amplification. Int J Otolaryngol 2014; 2014:641652. [PMID: 25045354 PMCID: PMC4086374 DOI: 10.1155/2014/641652] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Academic Contribution Register] [Received: 01/01/2014] [Revised: 05/01/2014] [Accepted: 05/14/2014] [Indexed: 01/19/2023] Open
Abstract
Hearing aids are used to improve sound audibility for people with hearing loss, but the ability to make use of the amplified signal, especially in the presence of competing noise, can vary across people. Here we review how neuroscientists, clinicians, and engineers are using various types of physiological information to improve the design and use of hearing aids.
Collapse
|
9
|
Electroacoustic Comparison of Hearing Aid Output of Phonemes in Running Speech versus Isolation: Implications for Aided Cortical Auditory Evoked Potentials Testing. Int J Otolaryngol 2012; 2012:518202. [PMID: 23316236 PMCID: PMC3536429 DOI: 10.1155/2012/518202] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Academic Contribution Register] [Received: 09/30/2012] [Accepted: 11/28/2012] [Indexed: 11/17/2022] Open
Abstract
Background. Functioning of nonlinear hearing aids varies with characteristics of input stimuli. In the past decade, aided speech evoked cortical auditory evoked potentials (CAEPs) have been proposed for validation of hearing aid fittings. However, unlike in running speech, phonemes presented as stimuli during CAEP testing are preceded by silent intervals of over one second. Hence, the present study aimed to compare if hearing aids process phonemes similarly in running speech and in CAEP testing contexts. Method. A sample of ten hearing aids was used. Overall phoneme level and phoneme onset level of eight phonemes in both contexts were compared at three input levels representing conversational speech levels. Results. Differences of over 3 dB between the two contexts were noted in one-fourth of the observations measuring overall phoneme levels and in one-third of the observations measuring phoneme onset level. In a majority of these differences, output levels of phonemes were higher in the running speech context. These differences varied across hearing aids. Conclusion. Lower output levels in the isolation context may have implications for calibration and estimation of audibility based on CAEPs. The variability across hearing aids observed could make it challenging to predict differences on an individual basis.
Collapse
|
10
|
The Influence of Audiovisual Ceiling Performance on the Relationship Between Reverberation and Directional Benefit. Ear Hear 2012; 33:604-14. [DOI: 10.1097/aud.0b013e31825641e4] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Academic Contribution Register] [Indexed: 11/25/2022]
|
11
|
Abstract
OBJECTIVES The current study was designed to see how hearing-impaired individuals judge level differences between speech sounds with and without hearing amplification. It was hypothesized that hearing aid compression should adversely affect the user's ability to judge level differences. DESIGN Thirty-eight hearing-impaired participants performed an adaptive tracking procedure to determine their level-discrimination thresholds for different word and sentence tokens, as well as speech-spectrum noise, with and without their hearing aids. Eight normal-hearing participants performed the same task for comparison. RESULTS Level discrimination for different word and sentence tokens was more difficult than the discrimination of stationary noises. Word level discrimination was significantly more difficult than sentence level discrimination. There were no significant differences, however, between mean performance with and without hearing aids and no correlations between performance and various hearing aid measurements. CONCLUSIONS There is a clear difficulty in judging the level differences between words or sentences relative to differences between broadband noises, but this difficulty was found for both hearing-impaired and normal-hearing individuals and had no relation to hearing aid compression measures. The lack of a clear adverse effect of hearing aid compression on level discrimination is suggested to be due to the low effective compression ratios of currently fit hearing aids.
Collapse
|
12
|
|