1
|
Jürgens T, Wesarg T, Oetting D, Jung L, Williges B. Spatial speech-in-noise performance in simulated single-sided deaf and bimodal cochlear implant users in comparison with real patients. Int J Audiol 2023; 62:30-43. [PMID: 34962428 DOI: 10.1080/14992027.2021.2015633] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2021] [Revised: 11/30/2021] [Accepted: 12/03/2021] [Indexed: 01/05/2023]
Abstract
OBJECTIVE Speech reception thresholds (SRTs) in spatial scenarios were measured in simulated cochlear implant (CI) listeners with either contralateral normal hearing, or aided hearing impairment (bimodal), and compared to SRTs of real patients, who were measured using the exact same paradigm, to assess goodness of simulation. DESIGN CI listening was simulated using a vocoder incorporating actual CI signal processing and physiologic details of electric stimulation on one side. Unprocessed signals or simulation of aided moderate or profound hearing impairment was used contralaterally. Three spatial speech-in-noise scenarios were tested using virtual acoustics to assess spatial release from masking (SRM) and combined benefit. STUDY SAMPLE Eleven normal-hearing listeners participated in the experiment. RESULTS For contralateral normal and aided moderately impaired hearing, bilaterally assessed SRTs were not statistically different from unilateral SRTs of the better ear, indicating "better-ear-listening". Combined benefit was only found for contralateral profound impaired hearing. As in patients, SRM was highest for contralateral normal hearing and decreased systematically with more severe simulated impairment. Comparison to actual patients showed good reproduction of SRTs, SRM, and better-ear-listening. CONCLUSIONS The simulations reproduced better-ear-listening as in patients and suggest that combined benefit in spatial scenes predominantly occurs when both ears show poor speech-in-noise performance.
Collapse
Affiliation(s)
- Tim Jürgens
- Institute of Acoustics, University of Applied Sciences Lübeck, Lübeck, Germany
- Medical Physics and Cluster of Excellence "Hearing4all", Carl-von-Ossietzky University, Oldenburg, Germany
| | - Thomas Wesarg
- Faculty of Medicine, Department of Otorhinolaryngology - Head and Neck Surgery, Medical Center, University of Freiburg, Freiburg, Germany
| | | | - Lorenz Jung
- Faculty of Medicine, Department of Otorhinolaryngology - Head and Neck Surgery, Medical Center, University of Freiburg, Freiburg, Germany
| | - Ben Williges
- Medical Physics and Cluster of Excellence "Hearing4all", Carl-von-Ossietzky University, Oldenburg, Germany
- SOUND Lab, Cambridge Hearing Group, Department of Clinical Neurosciences, University of Cambridge, Cambridge, UK
| |
Collapse
|
2
|
Modelling speech reception thresholds and their improvements due to spatial noise reduction algorithms in bimodal cochlear implant users. Hear Res 2022; 420:108507. [PMID: 35484022 PMCID: PMC9188268 DOI: 10.1016/j.heares.2022.108507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/19/2021] [Revised: 04/05/2022] [Accepted: 04/07/2022] [Indexed: 11/22/2022]
Abstract
This paper compares two modelling approaches to predict the speech recognition ability of bimodal CI users and the benefit of using beamformers. The modelling approaches vary in computational complexity and fitting requirements. A complex cafeteria spatial scenario with three localized single noise source scenario and a diffuse multi-talker babble noise is used. The automatic speech recognizer is more accurate across the different spatial scenarios and noise types and requires less fitting compared to the statistical modelling approach.
Spatial noise reduction algorithms (“beamformers”) can considerably improve speech reception thresholds (SRTs) for bimodal cochlear implant (CI) users. The goal of this study was to model SRTs and SRT-benefit due to beamformers for bimodal CI users. Two existing model approaches varying in computational complexity and binaural processing assumption were compared: (i) the framework of auditory discrimination experiments (FADE) and (ii) the binaural speech intelligibility model (BSIM), both with CI and aided hearing-impaired front-ends. The exact same acoustic scenarios, and open-access beamformers as in the comparison clinical study Zedan et al. (2021) were used to quantify goodness of prediction. FADE was capable of modeling SRTs ab-initio, i.e., no calibration of the model was necessary to achieve high correlations and low root-mean square errors (RMSE) to both, measured SRTs (r = 0.85, RMSE = 2.8 dB) and to measured SRT-benefits (r = 0.96). BSIM achieved somewhat poorer predictions to both, measured SRTs (r = 0.78, RMSE = 6.7 dB) and to measured SRT-benefits (r = 0.91) and needs to be calibrated for matching average SRTs in one condition. Greatest deviations in predictions of BSIM were observed in diffuse multi-talker babble noise, which were not found with FADE. SRT-benefit predictions of both models were similar to instrumental signal-to-noise ratio (iSNR) improvements due to the beamformers. This indicates that FADE is preferrable for modeling absolute SRTs. However, for prediction of SRT-benefit due to spatial noise reduction algorithms in bimodal CI users, the average iSNR is a much simpler approach with similar performance.
Collapse
|
3
|
Hauth CF, Berning SC, Kollmeier B, Brand T. Modeling Binaural Unmasking of Speech Using a Blind Binaural Processing Stage. Trends Hear 2020; 24:2331216520975630. [PMID: 33305690 PMCID: PMC7734536 DOI: 10.1177/2331216520975630] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The equalization cancellation model is often used to predict the binaural masking level difference. Previously its application to speech in noise has required separate knowledge about the speech and noise signals to maximize the signal-to-noise ratio (SNR). Here, a novel, blind equalization cancellation model is introduced that can use the mixed signals. This approach does not require any assumptions about particular sound source directions. It uses different strategies for positive and negative SNRs, with the switching between the two steered by a blind decision stage utilizing modulation cues. The output of the model is a single-channel signal with enhanced SNR, which we analyzed using the speech intelligibility index to compare speech intelligibility predictions. In a first experiment, the model was tested on experimental data obtained in a scenario with spatially separated target and masker signals. Predicted speech recognition thresholds were in good agreement with measured speech recognition thresholds with a root mean square error less than 1 dB. A second experiment investigated signals at positive SNRs, which was achieved using time compressed and low-pass filtered speech. The results demonstrated that binaural unmasking of speech occurs at positive SNRs and that the modulation-based switching strategy can predict the experimental results.
Collapse
Affiliation(s)
- Christopher F Hauth
- Medizinische Physik and Cluster of Excellence Hearing4All Carl-von-Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Simon C Berning
- Medizinische Physik and Cluster of Excellence Hearing4All Carl-von-Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Birger Kollmeier
- Medizinische Physik and Cluster of Excellence Hearing4All Carl-von-Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Thomas Brand
- Medizinische Physik and Cluster of Excellence Hearing4All Carl-von-Ossietzky Universität Oldenburg, Oldenburg, Germany
| |
Collapse
|
4
|
Vicente T, Lavandier M. Further validation of a binaural model predicting speech intelligibility against envelope-modulated noises. Hear Res 2020; 390:107937. [PMID: 32192940 DOI: 10.1016/j.heares.2020.107937] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/21/2019] [Revised: 02/24/2020] [Accepted: 02/28/2020] [Indexed: 10/24/2022]
Abstract
Collin and Lavandier [J. Acoust. Soc. Am. 134, 1146-1159 (2013)] proposed a binaural model predicting speech intelligibility against envelope-modulated noises, evaluated in 24 acoustic conditions, involving similar masker types. The aim of the present study was to test the model robustness modeling 80 additional conditions, and evaluate the influence of its parameters using an approach inspired by a variance-based sensitivity analysis. First, the data from four experiments from the literature and one specifically designed for the present study were used to evaluate the prediction performance of the model, investigate potential interactions between its parameters, and define their values leading to the best predictions. A revision of the model allowed to account for binaural sluggishness. Finally, the optimized model was tested on an additional dataset not used to define its parameters. Overall, one hundred conditions split into six experiments were modeled. Correlation between data and predictions ranged from 0.85 to 0.96 across experiments, and mean absolute prediction errors were between 0.5 and 1.4 dB.
Collapse
Affiliation(s)
- Thibault Vicente
- Univ Lyon, ENTPE, Laboratoire Génie Civil et Bâtiment, Rue Maurice Audin, 69518, Vaulx-en-Velin Cedex, France.
| | - Mathieu Lavandier
- Univ Lyon, ENTPE, Laboratoire Génie Civil et Bâtiment, Rue Maurice Audin, 69518, Vaulx-en-Velin Cedex, France
| |
Collapse
|
5
|
Rennies J, Warzybok A, Brand T, Kollmeier B. Measurement and Prediction of Binaural-Temporal Integration of Speech Reflections. Trends Hear 2019; 23:2331216519854267. [PMID: 31234732 PMCID: PMC6593929 DOI: 10.1177/2331216519854267] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
For speech intelligibility in rooms, the temporal integration of speech reflections is typically modeled by separating the room impulse response (RIR) into an early (assumed beneficial for speech intelligibility) and a late part (assumed detrimental). This concept was challenged in this study by employing binaural RIRs with systematically varied interaural phase differences (IPDs) and amplitude of the direct sound and a variable number of reflections delayed by up to 200 ms. Speech recognition thresholds in stationary noise were measured in normal-hearing listeners for 86 conditions. The data showed that direct sound and one or several early speech reflections could be perfectly integrated when they had the same IPD. Early reflections with the same IPD as the noise (but not as the direct sound) could not be perfectly integrated with the direct sound. All conditions in which the dominant speech information was within the early RIR components could be well predicted by a binaural speech intelligibility model using classic early/late separation. In contrast, when amplitude or IPD favored late RIR components, listeners appeared to be capable of focusing on these components rather than on the precedent direct sound. This could not be modeled by an early/late separation window but required a temporal integration window that can be flexibly shifted along the RIR.
Collapse
Affiliation(s)
- Jan Rennies
- 1 Department of Speech, Language and Hearing Sciences, Boston University, Boston, MA, USA.,2 Project Group Hearing, Speech and Audio Technology, Fraunhofer Institute for Digital Media Technology IDMT, Cluster of Excellence Hearing4all, Oldenburg, Germany
| | - Anna Warzybok
- 3 Medical Physics Group, Department of Medical Physics and Acoustics, Cluster of Excellence Hearing4all, University of Oldenburg, Germany
| | - Thomas Brand
- 3 Medical Physics Group, Department of Medical Physics and Acoustics, Cluster of Excellence Hearing4all, University of Oldenburg, Germany
| | - Birger Kollmeier
- 2 Project Group Hearing, Speech and Audio Technology, Fraunhofer Institute for Digital Media Technology IDMT, Cluster of Excellence Hearing4all, Oldenburg, Germany.,3 Medical Physics Group, Department of Medical Physics and Acoustics, Cluster of Excellence Hearing4all, University of Oldenburg, Germany
| |
Collapse
|
6
|
Hauth CF, Brand T. Modeling Sluggishness in Binaural Unmasking of Speech for Maskers With Time-Varying Interaural Phase Differences. Trends Hear 2019; 22:2331216517753547. [PMID: 29338577 PMCID: PMC5774735 DOI: 10.1177/2331216517753547] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
In studies investigating binaural processing in human listeners, relatively long and task-dependent time constants of a binaural window ranging from 10 ms to 250 ms have been observed. Such time constants are often thought to reflect “binaural sluggishness.” In this study, the effect of binaural sluggishness on binaural unmasking of speech in stationary speech-shaped noise is investigated in 10 listeners with normal hearing. In order to design a masking signal with temporally varying binaural cues, the interaural phase difference of the noise was modulated sinusoidally with frequencies ranging from 0.25 Hz to 64 Hz. The lowest, that is the best, speech reception thresholds (SRTs) were observed for the lowest modulation frequency. SRTs increased with increasing modulation frequency up to 4 Hz. For higher modulation frequencies, SRTs remained constant in the range of 1 dB to 1.5 dB below the SRT determined in the diotic situation. The outcome of the experiment was simulated using a short-term binaural speech intelligibility model, which combines an equalization–cancellation (EC) model with the speech intelligibility index. This model segments the incoming signal into 23.2-ms time frames in order to predict release from masking in modulated noises. In order to predict the results from this study, the model required a further time constant applied to the EC mechanism representing binaural sluggishness. The best agreement with perceptual data was achieved using a temporal window of 200 ms in the EC mechanism.
Collapse
Affiliation(s)
- Christopher F Hauth
- 1 Medizinische Physik, Carl von Ossietzky Universität, Oldenburg, Germany.,2 Cluster of Excellence Hearing4All, Carl von Ossietzky Universität, Oldenburg, Germany
| | - Thomas Brand
- 1 Medizinische Physik, Carl von Ossietzky Universität, Oldenburg, Germany.,2 Cluster of Excellence Hearing4All, Carl von Ossietzky Universität, Oldenburg, Germany
| |
Collapse
|
7
|
van Schoonhoven J, Rhebergen KS, Dreschler WA. The Extended Speech Transmission Index: Predicting speech intelligibility in fluctuating noise and reverberant rooms. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:1178. [PMID: 31067918 DOI: 10.1121/1.5092204] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/25/2018] [Accepted: 02/07/2019] [Indexed: 06/09/2023]
Abstract
The Speech Transmission Index (STI) is used to predict speech intelligibility in noise and reverberant environments. However, measurements and predictions in fluctuating noises lead to inaccuracies. In the current paper, the Extended Speech Transmission Index (ESTI) is presented in order to deal with these shortcomings. Speech intelligibility in normally hearing subjects was measured using stationary and fluctuating maskers. These results served to optimize model parameters. Data from the literature were then used to verify the ESTI-model. Model outcomes were accurate for stationary maskers, maskers with artificial fluctuations, and maskers with real life non-speech modulations. Maskers with speech-like characteristics introduced systematic errors in the model outcomes, probably due to a combination of modulation masking, context effects, and informational masking.
Collapse
Affiliation(s)
- Jelmer van Schoonhoven
- Department of Clinical and Experimental Audiology, Academic Medical Center, Meibergdreef 9, 1105 AZ Amsterdam, The Netherlands
| | - Koenraad S Rhebergen
- Department of Otorhinolaryngology and Head and Neck Surgery, Rudolf Magnus Institute of Neuroscience, University Medical Center, Utrecht, The Netherlands
| | - Wouter A Dreschler
- Department of Clinical and Experimental Audiology, Academic Medical Center, Meibergdreef 9, 1105 AZ Amsterdam, The Netherlands
| |
Collapse
|
8
|
Sheffield SW, Romigh GD, Zurek PM, Bernstein JGW, Brungart DS. A method for degrading sound localization while preserving binaural advantages for speech reception in noise. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:1129. [PMID: 30823825 DOI: 10.1121/1.5090494] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2018] [Accepted: 01/25/2019] [Indexed: 06/09/2023]
Abstract
This study developed and tested a real-time processing algorithm designed to degrade sound localization (LocDeg algorithm) without affecting binaural benefits for speech reception in noise. Input signals were divided into eight frequency channels. The odd-numbered channels were mixed between the ears to confuse the direction of interaural cues while preserving interaural cues in the even-numbered channels. The LocDeg algorithm was evaluated for normal-hearing listeners performing sound localization and speech-reception tasks. Results showed that the LocDeg algorithm successfully degraded sound-localization performance without affecting speech-reception performance or spatial release from masking for speech in noise. The LocDeg algorithm did, however, degrade speech-reception performance in a task involving spatially separated talkers in a multi-talker environment, which is thought to depend on differences in perceived spatial location of concurrent talkers. This LocDeg algorithm could be a valuable tool for isolating the importance of sound-localization ability from other binaural benefits in real-world environments.
Collapse
Affiliation(s)
- Sterling W Sheffield
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, 4954 North Palmer Road, Bethesda, Maryland 20889, USA
| | - Griffin D Romigh
- United States Air Force Research Laboratory, 5135 Pearson Road, Building 10, Wright Patterson Air Force Base, Ohio 45433, USA
| | - Patrick M Zurek
- Sensimetrics Corporation, 110 Winn Street, Suite 205A, Woburn, Massachusetts 01801, USA
| | - Joshua G W Bernstein
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, 4954 North Palmer Road, Bethesda, Maryland 20889, USA
| | - Douglas S Brungart
- National Military Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, 4954 North Palmer Road, Bethesda, Maryland 20889, USA
| |
Collapse
|
9
|
Verhey JL, van de Par S. Binaural frequency selectivity in humans. Eur J Neurosci 2018; 51:1179-1190. [PMID: 29359360 DOI: 10.1111/ejn.13837] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2017] [Revised: 01/09/2018] [Accepted: 01/10/2018] [Indexed: 11/30/2022]
Abstract
Several behavioural studies in humans have shown that listening to sounds with two ears that is binaural hearing, provides the human auditory system with extra information on the sound source that is not available when sounds are only perceived through one ear that is monaurally. Binaural processing involves the analysis of phase and level differences between the two ear signals. As monaural cochlea processing (in each ear) precedes the neural stages responsible for binaural processing properties it is reasonable to assume that properties of the cochlea may also be observed in binaural processing. A main characteristic of cochlea processing is its frequency selectivity. In psychoacoustics, there is an ongoing discussion on the frequency selectivity of the binaural auditory system. While some psychoacoustic experiments seem to indicate poorer frequency selectivity of the binaural system than that of the monaural processing others seem to indicate the same frequency selectivity for monaural and binaural processing. This study provides an overview of these seemingly controversial results and the different explanations that were provided to account for the different results.
Collapse
Affiliation(s)
- Jesko L Verhey
- Department of Experimental Audiology, Otto von Guericke University Magdeburg, Leipziger Str. 44, 39120, Magdeburg, Germany
| | - Steven van de Par
- Acoustic Group, Cluster of Excellence Hearing4All, Carl von Ossietzky University Oldenburg, Oldenburg, Germany
| |
Collapse
|
10
|
Kidd G, Colburn HS. Informational Masking in Speech Recognition. SPRINGER HANDBOOK OF AUDITORY RESEARCH 2017. [DOI: 10.1007/978-3-319-51662-2_4] [Citation(s) in RCA: 52] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
|
11
|
Wan R, Durlach NI, Colburn HS. Application of a short-time version of the Equalization-Cancellation model to speech intelligibility experiments with speech maskers. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 136:768-76. [PMID: 25096111 PMCID: PMC4144180 DOI: 10.1121/1.4884767] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2013] [Revised: 06/03/2014] [Accepted: 06/06/2014] [Indexed: 05/25/2023]
Abstract
A short-time-processing version of the Equalization-Cancellation (EC) model of binaural processing is described and applied to speech intelligibility tasks in the presence of multiple maskers, including multiple speech maskers. This short-time EC model, called the STEC model, extends the model described by Wan et al. [J. Acoust. Soc. Am. 128, 3678-3690 (2010)] to allow the EC model's equalization parameters τ and α to be adjusted as a function of time, resulting in improved masker cancellation when the dominant masker location varies in time. Using the Speech Intelligibility Index, the STEC model is applied to speech intelligibility with maskers that vary in number, type, and spatial arrangements. Most notably, when maskers are located on opposite sides of the target, this STEC model predicts improved thresholds when the maskers are modulated independently with speech-envelope modulators; this includes the most relevant case of independent speech maskers. The STEC model describes the spatial dependence of the speech reception threshold with speech maskers better than the steady-state model. Predictions are also improved for independently speech-modulated noise maskers but are poorer for reversed-speech maskers. In general, short-term processing is useful, but much remains to be done in the complex task of understanding speech in speech maskers.
Collapse
Affiliation(s)
- Rui Wan
- Hearing Research Center and Department of Biomedical Engineering, Boston University, Boston, Massachusetts 02215
| | - Nathaniel I Durlach
- Hearing Research Center and Department of Biomedical Engineering, Boston University, Boston, Massachusetts 02215
| | - H Steven Colburn
- Hearing Research Center and Department of Biomedical Engineering, Boston University, Boston, Massachusetts 02215
| |
Collapse
|
12
|
Cosentino S, Marquardt T, McAlpine D, Culling JF, Falk TH. A model that predicts the binaural advantage to speech intelligibility from the mixed target and interferer signals. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 135:796-807. [PMID: 25234888 DOI: 10.1121/1.4861239] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
A model is presented that predicts the binaural advantage to speech intelligibility by analyzing the right and left recordings at the two ears containing mixed target and interferer signals. This auditory-inspired model implements an equalization-cancellation stage to predict the binaural unmasking (BU) component, in conjunction with a modulation-frequency estimation block to estimate the "better ear" effect (BE) component of the binaural advantage. The model's performance was compared to experimental data obtained under anechoic and reverberant conditions using a single speech-shaped noise interferer paradigm. The internal BU and BE components were compared to those of the speech intelligibility model recently proposed by Lavandier et al. [J. Acoust. Soc. Am. 131, 218-231 (2012)], which requires separate inputs for target and interferer. The data indicate that the proposed model provides comparably good predictions from a mixed-signals input under both anechoic and reverberant conditions.
Collapse
Affiliation(s)
| | | | - David McAlpine
- Ear Institute, University College London, London, United Kingdom
| | - John F Culling
- School of Psychology, Cardiff University, Tower Building, Cardiff, United Kingdom
| | - Tiago H Falk
- Institut National de la Recherche Scientifique-Eńergie Matériaux Télécommunications, University of Québec, Montreal, Canada
| |
Collapse
|
13
|
Rothpletz AM, Wightman FL, Kistler DJ. Informational masking and spatial hearing in listeners with and without unilateral hearing loss. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2012; 55:511-531. [PMID: 22215037 PMCID: PMC3320681 DOI: 10.1044/1092-4388(2011/10-0205)] [Citation(s) in RCA: 45] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
PURPOSE This study assessed selective listening for speech in individuals with and without unilateral hearing loss (UHL) and the potential relationship between spatial release from informational masking and localization ability in listeners with UHL. METHOD Twelve adults with UHL and 12 normal-hearing controls completed a series of monaural and binaural speech tasks that were designed to measure informational masking. They also completed a horizontal localization task. RESULTS Monaural performance by participants with UHL was comparable to that of normal-hearing participants. Unlike the normal-hearing participants, the participants with UHL did not exhibit a true spatial release from informational masking. Rather, their performance could be predicted by head shadow effects. Performance among participants with UHL in the localization task was quite variable, with some showing near-normal abilities and others demonstrating no localization ability. CONCLUSION Individuals with UHL did not show deficits in all listening situations but were at a significant disadvantage when listening to speech in environments where normal-hearing listeners benefit from spatial separation between target and masker. This inability to capitalize on spatial cues for selective listening does not appear to be related to localization ability.
Collapse
|
14
|
Jelfs S, Culling JF, Lavandier M. Revision and validation of a binaural model for speech intelligibility in noise. Hear Res 2011; 275:96-104. [DOI: 10.1016/j.heares.2010.12.005] [Citation(s) in RCA: 23] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/27/2010] [Revised: 12/01/2010] [Accepted: 12/06/2010] [Indexed: 10/18/2022]
|
15
|
Beutelmann R, Brand T, Kollmeier B. Revision, extension, and evaluation of a binaural speech intelligibility model. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 127:2479-97. [PMID: 20370031 DOI: 10.1121/1.3295575] [Citation(s) in RCA: 52] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
This study presents revision, extension, and evaluation of a binaural speech intelligibility model (Beutelmann, R., and Brand, T. (2006). J. Acoust. Soc. Am. 120, 331-342) that yields accurate predictions of speech reception thresholds (SRTs) in the presence of a stationary noise source at arbitrary azimuths and in different rooms. The modified model is based on an analytical expression of binaural unmasking for arbitrary input signals and is computationally more efficient, while maintaining the prediction quality of the original model. An extension for nonstationary interferers was realized by applying the model to short time frames of the input signals and averaging over the predicted SRT results. Binaural SRTs from 8 normal-hearing and 12 hearing-impaired subjects, incorporating all combinations of four rooms, three source setups, and three noise types were measured and compared to the model's predictions. Depending on the noise type, the parametric correlation coefficients between observed and predicted SRTs were 0.80-0.93 for normal-hearing subjects and 0.59-0.80 for hearing-impaired subjects. The mean absolute prediction error was 3 dB for the mean normal-hearing data and 4 dB for the individual hearing-impaired data. 70% of the variance of the SRTs of hearing-impaired subjects could be explained by the model, which is based only on the audiogram.
Collapse
Affiliation(s)
- Rainer Beutelmann
- Medizinische Physik, Carl-von-Ossietzky-Universitat Oldenburg, 26111 Oldenburg, Germany.
| | | | | |
Collapse
|