1
|
Thaler L, Castillo-Serrano JG, Kish D, Norman LJ. Effects of type of emission and masking sound, and their spatial correspondence, on blind and sighted people's ability to echolocate. Neuropsychologia 2024; 196:108822. [PMID: 38342179 DOI: 10.1016/j.neuropsychologia.2024.108822] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Revised: 01/30/2024] [Accepted: 02/08/2024] [Indexed: 02/13/2024]
Abstract
Ambient sound can mask acoustic signals. The current study addressed how echolocation in people is affected by masking sound, and the role played by type of sound and spatial (i.e. binaural) similarity. We also investigated the role played by blindness and long-term experience with echolocation, by testing echolocation experts, as well as blind and sighted people new to echolocation. Results were obtained in two echolocation tasks where participants listened to binaural recordings of echolocation and masking sounds, and either localized echoes in azimuth or discriminated echo audibility. Echolocation and masking sounds could be either clicks or broad band noise. An adaptive staircase method was used to adjust signal-to-noise ratios (SNRs) based on participants' responses. When target and masker had the same binaural cues (i.e. both were monoaural sounds), people performed better (i.e. had lower SNRs) when target and masker used different types of sound (e.g. clicks in noise-masker or noise in clicks-masker), as compared to when target and masker used the same type of sound (e.g. clicks in click-, or noise in noise-masker). A very different pattern of results was observed when masker and target differed in their binaural cues, in which case people always performed better when clicks were the masker, regardless of type of emission used. Further, direct comparison between conditions with and without binaural difference revealed binaural release from masking only when clicks were used as emissions and masker, but not otherwise (i.e. when noise was used as masker or emission). This suggests that echolocation with clicks or noise may differ in their sensitivity to binaural cues. We observed the same pattern of results for echolocation experts, and blind and sighted people new to echolocation, suggesting a limited role played by long-term experience or blindness. In addition to generating novel predictions for future work, the findings also inform instruction in echolocation for people who are blind or sighted.
Collapse
Affiliation(s)
- L Thaler
- Department of Psychology, Durham University, South Road, Durham, DH1 5AY, UK.
| | | | - D Kish
- World Access for the Blind, 1007 Marino Drive, Placentia, CA, 92870, USA
| | - L J Norman
- Department of Psychology, Durham University, South Road, Durham, DH1 5AY, UK
| |
Collapse
|
2
|
Oh Y, Friggle P, Kinder J, Tilbrook G, Bridges SE. Effects of presentation level on speech-on-speech masking by voice-gender difference and spatial separation between talkers. Front Neurosci 2023; 17:1282764. [PMID: 38192513 PMCID: PMC10773857 DOI: 10.3389/fnins.2023.1282764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 11/30/2023] [Indexed: 01/10/2024] Open
Abstract
Many previous studies have reported that speech segregation performance in multi-talker environments can be enhanced by two major acoustic cues: (1) voice-characteristic differences between talkers; (2) spatial separation between talkers. Here, the improvement they can provide for speech segregation is referred to as "release from masking." The goal of this study was to investigate how masking release performance with two cues is affected by various target presentation levels. Sixteen normal-hearing listeners participated in the speech recognition in noise experiment. Speech-on-speech masking performance was measured as the threshold target-to-masker ratio needed to understand a target talker in the presence of either same- or different-gender masker talkers to manipulate the voice-gender difference cue. These target-masker gender combinations were tested with five spatial configurations (maskers co-located or 15°, 30°, 45°, and 60° symmetrically spatially separated from the target) to manipulate the spatial separation cue. In addition, those conditions were repeated at three target presentation levels (30, 40, and 50 dB sensation levels). Results revealed that the amount of masking release by either voice-gender difference or spatial separation cues was significantly affected by the target level, especially at the small target-masker spatial separation (±15°). Further, the results showed that the intersection points between two masking release types (equal perceptual weighting) could be varied by the target levels. These findings suggest that the perceptual weighting of masking release from two cues is non-linearly related to the target levels. The target presentation level could be one major factor associated with masking release performance in normal-hearing listeners.
Collapse
Affiliation(s)
- Yonghee Oh
- Department of Otolaryngology-Head and Neck Surgery and Communicative Disorders, University of Louisville, Louisville, KY, United States
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville, FL, United States
| | - Phillip Friggle
- Department of Otolaryngology-Head and Neck Surgery and Communicative Disorders, University of Louisville, Louisville, KY, United States
| | - Josephine Kinder
- Department of Otolaryngology-Head and Neck Surgery and Communicative Disorders, University of Louisville, Louisville, KY, United States
| | - Grace Tilbrook
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville, FL, United States
| | - Sarah E. Bridges
- Department of Speech, Language, and Hearing Sciences, University of Florida, Gainesville, FL, United States
| |
Collapse
|
3
|
Wasiuk PA, Calandruccio L, Oleson JJ, Buss E. Predicting speech-in-speech recognition: Short-term audibility and spatial separation. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 154:1827-1837. [PMID: 37728286 DOI: 10.1121/10.0021069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/18/2023] [Accepted: 08/28/2023] [Indexed: 09/21/2023]
Abstract
Quantifying the factors that predict variability in speech-in-speech recognition represents a fundamental challenge in auditory science. Stimulus factors associated with energetic and informational masking (IM) modulate variability in speech-in-speech recognition, but energetic effects can be difficult to estimate in spectro-temporally dynamic speech maskers. The current experiment characterized the effects of short-term audibility and differences in target and masker location (or perceived location) on the horizontal plane for sentence recognition in two-talker speech. Thirty young adults with normal hearing (NH) participated. Speech reception thresholds and keyword recognition at a fixed signal-to-noise ratio (SNR) were measured in each spatial condition. Short-term audibility for each keyword was quantified using a glimpsing model. Results revealed that speech-in-speech recognition depended on the proportion of audible glimpses available in the target + masker keyword stimulus in each spatial condition, even across stimuli presented at a fixed global SNR. Short-term audibility requirements were greater for colocated than spatially separated speech-in-speech recognition, and keyword recognition improved more rapidly as a function of increases in target audibility with spatial separation. Results indicate that spatial cues enhance glimpsing efficiency in competing speech for young adults with NH and provide a quantitative framework for estimating IM for speech-in-speech recognition in different spatial configurations.
Collapse
Affiliation(s)
- Peter A Wasiuk
- Department of Communication Disorders, 493 Fitch Street, Southern Connecticut State University, New Haven, Connecticut 06515, USA
| | - Lauren Calandruccio
- Department of Psychological Sciences, 11635 Euclid Avenue, Case Western Reserve University, Cleveland, Ohio 44106, USA
| | - Jacob J Oleson
- Department of Biostatistics, 145 North Riverside Drive N300, College of Public Health, University of Iowa, Iowa City, Iowa 52242, USA
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, 170 Manning Drive, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA
| |
Collapse
|
4
|
Maillard E, Joyal M, Murray MM, Tremblay P. Are musical activities associated with enhanced speech perception in noise in adults? A systematic review and meta-analysis. CURRENT RESEARCH IN NEUROBIOLOGY 2023. [DOI: 10.1016/j.crneur.2023.100083] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023] Open
|
5
|
Lelo de Larrea-Mancera ES, Solís-Vivanco R, Sánchez-Jimenez Y, Coco L, Gallun FJ, Seitz AR. Development and validation of a Spanish-language spatial release from masking task in a Mexican population. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 153:316. [PMID: 36732214 PMCID: PMC10162838 DOI: 10.1121/10.0016850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/07/2023]
Abstract
This study validates a new Spanish-language version of the Coordinate Response Measure (CRM) corpus using a well-established measure of spatial release from masking (SRM). Participants were 96 Spanish-speaking young adults without hearing complaints in Mexico City. To present the Spanish-language SRM test, we created new recordings of the CRM with Spanish-language Translations and updated the freely available app (PART; https://ucrbraingamecenter.github.io/PART_Utilities/) to present materials in Spanish. In addition to SRM, we collected baseline data on a battery of non-speech auditory assessments, including detection of frequency modulations, temporal gaps, and modulated broadband noise in the temporal, spectral, and spectrotemporal domains. Data demonstrate that the newly developed speech and non-speech tasks show similar reliability to an earlier report in English-speaking populations. This study demonstrates an approach by which auditory assessment for clinical and basic research can be extended to Spanish-speaking populations for whom testing platforms are not currently available.
Collapse
Affiliation(s)
| | - Rodolfo Solís-Vivanco
- Laboratory of Cognitive and Clinical Neurophysiology, Instituto Nacional de Neurología y Neurocirugía Manuel Velasco Suárez (INNNMVS), Avenue Insurgentes Sur 3877, La Fama, Tlalpan, Mexico City, CDMX 14269, Mexico
| | | | - Laura Coco
- Department of Otolaryngology, Oregon Health & Science University, Portland, Oregon 97239, USA
| | - Frederick J Gallun
- Department of Otolaryngology, Oregon Health & Science University, Portland, Oregon 97239, USA
| | - Aaron R Seitz
- Department of Psychology, University of California, 900 University Avenue, Riverside, California 92507, USA
| |
Collapse
|
6
|
Ozmeral EJ, Higgins NC. Defining functional spatial boundaries using a spatial release from masking task. JASA EXPRESS LETTERS 2022; 2:124402. [PMID: 36586966 PMCID: PMC9720634 DOI: 10.1121/10.0015356] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/28/2022] [Accepted: 11/11/2022] [Indexed: 06/17/2023]
Abstract
The classic spatial release from masking (SRM) task measures speech recognition thresholds for discrete separation angles between a target and masker. Alternatively, this study used a modified SRM task that adaptively measured the spatial-separation angle needed between a continuous male target stream (speech with digits) and two female masker streams to achieve a specific SRM. On average, 20 young normal-hearing listeners needed less spatial separation for 6 dB release than 9 dB release, and the presence of background babble reduced across-listener variability on the paradigm. Future work is needed to better understand the psychometric properties of this adaptive procedure.
Collapse
Affiliation(s)
- Erol J Ozmeral
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA ,
| | - Nathan C Higgins
- Department of Communication Sciences and Disorders, University of South Florida, Tampa, Florida 33620, USA ,
| |
Collapse
|
7
|
Stenbäck V, Marsja E, Hällgren M, Lyxell B, Larsby B. Informational Masking and Listening Effort in Speech Recognition in Noise: The Role of Working Memory Capacity and Inhibitory Control in Older Adults With and Without Hearing Impairment. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:4417-4428. [PMID: 36283680 DOI: 10.1044/2022_jslhr-21-00674] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
PURPOSE The study aimed to assess the relationship between (a) speech recognition in noise, mask type, working memory capacity (WMC), and inhibitory control and (b) self-rated listening effort, speech material, and mask type, in older adults with and without hearing impairment. It was of special interest to assess the relationship between WMC, inhibitory control, and speech recognition in noise when informational maskers masked target speech. METHOD A mixed design was used. A group (N = 24) of older (Mage = 69.7 years) individuals with hearing impairment and a group of age normal-hearing adults (Mage = 59.3 years, SD = 6.5) participated in the study. The participants were presented with auditory tests in a sound-attenuated room and with cognitive tests in a quiet office. The participants were asked to rate listening effort after being presented with energetic and informational background maskers in two different speech materials used in this study (i.e., Hearing In Noise Test and Hagerman test). Linear mixed-effects models were set up to assess the effect of the two different speech materials, energetic and informational maskers, hearing ability, WMC, inhibitory control, and self-rated listening effort. RESULTS Results showed that WMC and inhibitory control were of importance for speech recognition in noise, even when controlling for pure-tone average 4 hearing thresholds and age, when the maskers were informational. Concerning listening effort, on the other hand, the results suggest that hearing ability, but not cognitive abilities, is important for self-rated listening effort in speech recognition in noise. CONCLUSIONS Speech-in-noise recognition is more dependent on WMC for older adults in informational maskers than in energetic maskers. Hearing ability is a stronger predictor than cognition for self-rated listening effort. SUPPLEMENTAL MATERIAL https://doi.org/10.23641/asha.21357648.
Collapse
Affiliation(s)
- Victoria Stenbäck
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Division of Education, Teaching and Learning, Department of Behavioural Sciences and Learning, Linköping University, Sweden
| | - Erik Marsja
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
| | - Mathias Hällgren
- Department of Otorhinolaryngology in Östergötland and Department of Biomedical and Clinical Sciences, Linköping University, Sweden
| | - Björn Lyxell
- Disability Research Division, Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Department of Special Needs Education, University of Oslo, Norway
| | - Birgitta Larsby
- Department of Otorhinolaryngology in Östergötland and Department of Biomedical and Clinical Sciences, Linköping University, Sweden
| |
Collapse
|
8
|
Napierała M, Suwalska A, Pucher B, Rybakowski JK. Speech Understanding in Manic and Depressive Episodes of Mood Disorders. J Neuropsychiatry Clin Neurosci 2022; 34:414-421. [PMID: 35414193 DOI: 10.1176/appi.neuropsych.21050125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
OBJECTIVE The aim of this study was to assess the perception of speech in adverse acoustic conditions during manic and depressive episodes of mood disorders. METHODS Forty-three patients with bipolar disorder (mania, N=20; depression, N=23) and 32 patients with unipolar depression were included for analyses. Thirty-five participants served as the control group. The study of speech understanding was carried out using the Polish Sentence Matrix Test, allowing for the determination of the speech reception threshold (SRT). The test was performed in the clinical groups both during an acute episode and remission; during remission, patients underwent audiometric evaluation. RESULTS Compared with control subjects, patients with mood disorders had worse speech understanding (higher SRT), regardless of the episode or remission. A manic episode in the course of bipolar disorder was not associated with worse speech understanding compared with remission of mania. However, an episode of depression in the course of both bipolar disorder and unipolar depression was associated with worse speech understanding compared with remission of depression. In bipolar depression, this correlated with age, duration of the disorder, number of episodes, and number of hospitalizations, as well as in remission with age and duration of illness. In unipolar depression, poor speech understanding was more severe in individuals with hearing impairment. CONCLUSIONS These findings revealed that patients with mood disorders had impaired speech understanding, even while in remission, and manic episodes in the course of bipolar disorder were not associated with impaired speech understanding compared with mania remission.
Collapse
Affiliation(s)
- Maria Napierała
- Departments of Mental Health (Napierała, Suwalska), Adult Psychiatry (Napierała, Rybakowski), and Paediatric Otolaryngology (Pucher), Poznan University of Medical Sciences, Poznan, Poland
| | - Aleksandra Suwalska
- Departments of Mental Health (Napierała, Suwalska), Adult Psychiatry (Napierała, Rybakowski), and Paediatric Otolaryngology (Pucher), Poznan University of Medical Sciences, Poznan, Poland
| | - Beata Pucher
- Departments of Mental Health (Napierała, Suwalska), Adult Psychiatry (Napierała, Rybakowski), and Paediatric Otolaryngology (Pucher), Poznan University of Medical Sciences, Poznan, Poland
| | - Janusz K Rybakowski
- Departments of Mental Health (Napierała, Suwalska), Adult Psychiatry (Napierała, Rybakowski), and Paediatric Otolaryngology (Pucher), Poznan University of Medical Sciences, Poznan, Poland
| |
Collapse
|
9
|
Bestel J, Legris E, Rembaud F, Mom T, Galvin JJ. Speech understanding in diffuse steady noise in typically hearing and hard of hearing listeners. PLoS One 2022; 17:e0274435. [PMID: 36103551 PMCID: PMC9473430 DOI: 10.1371/journal.pone.0274435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2021] [Accepted: 08/29/2022] [Indexed: 12/02/2022] Open
Abstract
Spatial cues can facilitate segregation of target speech from maskers. However, in clinical practice, masked speech understanding is most often evaluated using co-located speech and maskers (i.e., without spatial cues). Many hearing aid centers in France are equipped with five-loudspeaker arrays, allowing masked speech understanding to be measured with spatial cues. It is unclear how hearing status may affect utilization of spatial cues to segregate speech and noise. In this study, speech reception thresholds (SRTs) for target speech in “diffuse noise” (target speech from 1 speaker, noise from the remaining 4 speakers) in 297 adult listeners across 9 Audilab hearing centers. Participants were categorized according to pure-tone-average (PTA) thresholds: typically-hearing (TH; ≤ 20 dB HL), mild hearing loss (Mild; >20 ≤ 40 dB HL), moderate hearing loss 1 (Mod-1; >40 ≤ 55 dB HL), and moderate hearing loss 2 (Mod-2; >55 ≤ 65 dB HL). All participants were tested without aided hearing. SRTs in diffuse noise were significantly correlated with PTA thresholds, age at testing, as well as word and phoneme recognition scores in quiet. Stepwise linear regression analysis showed that SRTs in diffuse noise were significantly predicted by a combination of PTA threshold and word recognition scores in quiet. SRTs were also measured in co-located and diffuse noise in 65 additional participants. SRTs were significantly lower in diffuse noise than in co-located noise only for the TH and Mild groups; masking release with diffuse noise (relative to co-located noise) was significant only for the TH group. The results are consistent with previous studies that found that hard of hearing listeners have greater difficulty using spatial cues to segregate competing speech. The data suggest that speech understanding in diffuse noise provides additional insight into difficulties that hard of hearing individuals experience in complex listening environments.
Collapse
Affiliation(s)
| | | | | | - Thierry Mom
- Centre Hospitalier Universitaire de Clermont-Ferrand, Clermont-Ferrand, France
| | - John J. Galvin
- University Hospital Center of Tours, Tours, France
- House Institute Foundation, Los Angeles, CA, United States of America
- * E-mail:
| |
Collapse
|
10
|
Mepham A, Bi Y, Mattys SL. The time-course of linguistic interference during native and non-native speech-in-speech listening. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:954. [PMID: 36050191 DOI: 10.1121/10.0013417] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Accepted: 07/20/2022] [Indexed: 06/15/2023]
Abstract
Recognizing speech in a noisy background is harder when the background is time-forward than for time-reversed speech, a masker direction effect, and harder when the masker is in a known rather than an unknown language, indicating linguistic interference. We examined the masker direction effect when the masker was a known vs unknown language and calculated performance over 50 trials to assess differential masker adaptation. In experiment 1, native English listeners transcribing English sentences showed a larger masker direction effect with English than Mandarin maskers. In experiment 2, Mandarin non-native speakers of English transcribing Mandarin sentences showed a mirror pattern. Both experiments thus support the target-masker linguistic similarity hypothesis, where interference is maximal when target and masker languages are the same. In experiment 3, Mandarin non-native speakers of English transcribing English sentences showed comparable results for English and Mandarin maskers. Non-native listening is therefore consistent with the known-language interference hypothesis, where interference is maximal when the masker language is known to the listener, whether or not it matches the target language. A trial-by-trial analysis showed that the masker direction effect increased over time during native listening but not during non-native listening. The results indicate different target-to-masker streaming strategies during native and non-native speech-in-speech listening.
Collapse
Affiliation(s)
- Alex Mepham
- Department of Psychology, University of York, Heslington, United Kingdom
| | - Yifei Bi
- College of Foreign Languages, University of Shanghai for Science and Technology, Shanghai, China
| | - Sven L Mattys
- Department of Psychology, University of York, Heslington, United Kingdom
| |
Collapse
|
11
|
Zobel BH, Freyman RL, Sanders LD. Spatial release from informational masking enhances the early cortical representation of speech sounds. AUDITORY PERCEPTION & COGNITION 2022; 5:211-237. [PMID: 36160272 PMCID: PMC9494573 DOI: 10.1080/25742442.2022.2088329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2021] [Accepted: 06/04/2022] [Indexed: 06/16/2023]
Abstract
INTRODUCTION Spatial separation between competing speech streams reduces their confusion (informational masking), improving speech processing under challenging listening conditions. The precise stages of auditory processing involved in this benefit are not fully understood. This study used event-related potentials to examine the processing of target speech under conditions of informational masking and its spatial release. METHODS Participants detected noise-vocoded target speech presented with two-talker noise-vocoded masking speech. In separate conditions, the same set of targets were spatially co-located with maskers to produce informational masking and spatially separated from maskers using a perceptual manipulation to release the informational masking. RESULTS An increase in N1 and P2 amplitude, consistent with cortical auditory evoked potentials, and a later sustained positivity (P300) were observed in response to target onsets only under conditions supporting release from informational masking. At target intensities above masking threshold in both spatial conditions, N1 and P2 latencies were shorter when targets and maskers were perceptually separated. DISCUSSION These results indicate that spatial release from informational masking benefits speech representation beginning in the early stages of auditory perception. Additionally, these results suggest that the auditory evoked potential itself may be heavily dependent upon how information is perceptually organized rather than physically organized.
Collapse
Affiliation(s)
- Benjamin H. Zobel
- Department of Psychological and Brain Sciences, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Richard L. Freyman
- Department of Communication Disorders, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| | - Lisa D. Sanders
- Department of Psychological and Brain Sciences, University of Massachusetts Amherst, Amherst, Massachusetts 01003
| |
Collapse
|
12
|
Effect of Masker Head Orientation, Listener Age, and Extended High-Frequency Sensitivity on Speech Recognition in Spatially Separated Speech. Ear Hear 2022; 43:90-100. [PMID: 34260434 PMCID: PMC8712343 DOI: 10.1097/aud.0000000000001081] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
Abstract
OBJECTIVES Masked speech recognition is typically assessed as though the target and background talkers are all directly facing the listener. However, background speech in natural environments is often produced by talkers facing other directions, and talker head orientation affects the spectral content of speech, particularly at the extended high frequencies (EHFs; >8 kHz). This study investigated the effect of masker head orientation and listeners' EHF sensitivity on speech-in-speech recognition and spatial release from masking in children and adults. DESIGN Participants were 5- to 7-year-olds (n = 15) and adults (n = 34), all with normal hearing up to 8 kHz and a range of EHF hearing thresholds. Speech reception thresholds (SRTs) were measured for target sentences recorded from a microphone directly in front of the talker's mouth and presented from a loudspeaker directly in front of the listener, simulating a target directly in front of and facing the listener. The maskers were two streams of concatenated words recorded from a microphone located at either 0° or 60° azimuth, simulating masker talkers facing the listener or facing away from the listener, respectively. Maskers were presented in one of three spatial conditions: co-located with the target, symmetrically separated on either side of the target (+54° and -54° on the horizontal plane), or asymmetrically separated to the right of the target (both +54° on the horizontal plane). RESULTS Performance was poorer for the facing than for the nonfacing masker head orientation. This benefit of the nonfacing masker head orientation, or head orientation release from masking (HORM), was largest under the co-located condition, but it was also observed for the symmetric and asymmetric masker spatial separation conditions. SRTs were positively correlated with the mean 16-kHz threshold across ears in adults for the nonfacing conditions but not for the facing masker conditions. In adults with normal EHF thresholds, the HORM was comparable in magnitude to the benefit of a symmetric spatial separation of the target and maskers. Although children benefited from the nonfacing masker head orientation, their HORM was reduced compared to adults with normal EHF thresholds. Spatial release from masking was comparable across age groups for symmetric masker placement, but it was larger in adults than children for the asymmetric masker. CONCLUSIONS Masker head orientation affects speech-in-speech recognition in children and adults, particularly those with normal EHF thresholds. This is important because masker talkers do not all face the listener under most natural listening conditions, and assuming a midline orientation would tend to overestimate the effect of spatial separation. The benefits associated with EHF audibility for speech-in-speech recognition may warrant clinical evaluation of thresholds above 8 kHz.
Collapse
|
13
|
Heeren J, Nuesse T, Latzel M, Holube I, Hohmann V, Wagener KC, Schulte M. The Concurrent OLSA Test: A Method for Speech Recognition in Multi-talker Situations at Fixed SNR. Trends Hear 2022; 26:23312165221108257. [PMID: 35702051 PMCID: PMC9208053 DOI: 10.1177/23312165221108257] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
A multi-talker paradigm is introduced that uses different attentional processes to adjust speech-recognition scores with the goal of conducting measurements at high signal-to-noise ratios (SNR). The basic idea is to simulate a group conversation with three talkers. Talkers alternately speak sentences of the German matrix test OLSA. Each time a sentence begins with the name “Kerstin” (call sign), the participant is addressed and instructed to repeat the last words of all sentences from that talker, until another talker begins a sentence with “Kerstin”. The alternation of the talkers is implemented with an adjustable overlap time that causes an overlap between the call sign “Kerstin” and the target words to be repeated. Thus, the two tasks of detecting “Kerstin” and repeating target words are to be done at the same time. The paradigm was tested with 22 young normal-hearing participants (YNH) for three overlap times (0.6 s, 0.8 s, 1.0 s). Results for these overlap times show significant differences, with median target word recognition scores of 88%, 82%, and 77%, respectively (including call-sign and dual-task effects). A comparison of the dual task with the corresponding single tasks suggests that the observed effects reflect an increased cognitive load.
Collapse
Affiliation(s)
- Jan Heeren
- Hörzentrum Oldenburg gGmbH, Oldenburg, Germany.,Cluster of Excellence Hearing4All, Oldenburg, Germany
| | - Theresa Nuesse
- Cluster of Excellence Hearing4All, Oldenburg, Germany.,Institute of Hearing Technology and Audiology, Jade University of Applied Sciences, Oldenburg, Germany
| | | | - Inga Holube
- Cluster of Excellence Hearing4All, Oldenburg, Germany.,Institute of Hearing Technology and Audiology, Jade University of Applied Sciences, Oldenburg, Germany
| | - Volker Hohmann
- Hörzentrum Oldenburg gGmbH, Oldenburg, Germany.,Cluster of Excellence Hearing4All, Oldenburg, Germany.,Auditory Signal Processing, Department of Medical Physics and Acoustics, University of Oldenburg, Oldenburg, Germany
| | - Kirsten C Wagener
- Hörzentrum Oldenburg gGmbH, Oldenburg, Germany.,Cluster of Excellence Hearing4All, Oldenburg, Germany
| | - Michael Schulte
- Hörzentrum Oldenburg gGmbH, Oldenburg, Germany.,Cluster of Excellence Hearing4All, Oldenburg, Germany
| |
Collapse
|
14
|
Zhang M, Denison RN, Pelli DG, Le TTC, Ihlefeld A. An auditory-visual tradeoff in susceptibility to clutter. Sci Rep 2021; 11:23540. [PMID: 34876580 PMCID: PMC8651672 DOI: 10.1038/s41598-021-00328-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Accepted: 08/12/2021] [Indexed: 01/13/2023] Open
Abstract
Sensory cortical mechanisms combine auditory or visual features into perceived objects. This is difficult in noisy or cluttered environments. Knowing that individuals vary greatly in their susceptibility to clutter, we wondered whether there might be a relation between an individual's auditory and visual susceptibilities to clutter. In auditory masking, background sound makes spoken words unrecognizable. When masking arises due to interference at central auditory processing stages, beyond the cochlea, it is called informational masking. A strikingly similar phenomenon in vision, called visual crowding, occurs when nearby clutter makes a target object unrecognizable, despite being resolved at the retina. We here compare susceptibilities to auditory informational masking and visual crowding in the same participants. Surprisingly, across participants, we find a negative correlation (R = -0.7) between susceptibility to informational masking and crowding: Participants who have low susceptibility to auditory clutter tend to have high susceptibility to visual clutter, and vice versa. This reveals a tradeoff in the brain between auditory and visual processing.
Collapse
Affiliation(s)
- Min Zhang
- grid.260896.30000 0001 2166 4955Department of Biomedical Engineering, New Jersey Institute of Technology, Newark, NJ USA ,grid.430387.b0000 0004 1936 8796Department of Biomedical Engineering, Rutgers New Jersey Medical School, Newark, NJ USA
| | - Rachel N Denison
- grid.189504.10000 0004 1936 7558Department of Psychology, Boston University, Boston, MA USA
| | - Denis G Pelli
- grid.137628.90000 0004 1936 8753Department of Psychology, New York University, New York, NY USA
| | - Thuy Tien C Le
- grid.260896.30000 0001 2166 4955Department of Biomedical Engineering, New Jersey Institute of Technology, Newark, NJ USA ,grid.430387.b0000 0004 1936 8796Department of Biomedical Engineering, Rutgers New Jersey Medical School, Newark, NJ USA
| | - Antje Ihlefeld
- Department of Biomedical Engineering, New Jersey Institute of Technology, Newark, NJ, USA.
| |
Collapse
|
15
|
Zhang M, Moncrieff D, Johnston D, Parfitt M, Auld R. A preliminary study on speech recognition in noise training for children with hearing loss. Int J Pediatr Otorhinolaryngol 2021; 149:110843. [PMID: 34340007 DOI: 10.1016/j.ijporl.2021.110843] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/28/2020] [Revised: 06/27/2021] [Accepted: 07/16/2021] [Indexed: 10/20/2022]
Abstract
PURPOSE The current study is a preliminary study to examine whether children with hearing loss would benefit from a speech recognition in noise training. METHODS Twenty-five children who wore hearing aids, cochlear implants, or bimodal devices from 4 to 12 years old participated in the study (experimental, n = 16; control, n = 9). The experimental group received a speech-in-noise training that took sixteen 15-min sessions spanning 8 to 12 weeks. The task involves recognizing monosyllabic target words and sentence keywords with various contextual cues in a multi-talker babble. The target stimuli were spoken by two females and fixed at 65 dB SPL throughout the training while the masker varied adaptively. Pre- and post-training tests measured the speech recognition thresholds of monosyllabic words and sentences spoken by two males in the babble noise. The test targets were presented at 55, 65, and 80 dB SPL. RESULTS The experimental group improved for word and sentence recognition in noise after training (Mean Difference = 2.4-2.5 dB, 2.7-4.2 dB, respectively). Training benefits were observed at trained (65 dB SPL) and untrained levels (55 and 80 dB SPL). The amount of post-training improvement was comparable between children using hearing aids and cochlear implants. CONCLUSIONS This preliminary study showed that children with hearing loss could benefit from a speech recognition in noise training that may fit into the children's school schedules. Training at a conversational level (65 dB SPL) transfers the benefit to levels 10-15 dB softer or louder. Training with female target talkers transfers the benefit to male target talkers. Overall, speech in noise training brings practical benefits for school-age children with hearing loss.
Collapse
Affiliation(s)
- Mengchao Zhang
- Department of Communication Science and Disorders, University of Pittsburgh, 6035 Forbes Tower, Pittsburgh, PA, 15260, USA.
| | - Deborah Moncrieff
- School of Communication Sciences and Disorders, University of Memphis, 4055 N. Park Loop, Memphis, TN, 38152, USA
| | - Deborrah Johnston
- DePaul School for Hearing and Speech, 6202 Alder St, Pittsburgh, PA, 15206, USA
| | - Michelle Parfitt
- DePaul School for Hearing and Speech, 6202 Alder St, Pittsburgh, PA, 15206, USA
| | - Ruth Auld
- DePaul School for Hearing and Speech, 6202 Alder St, Pittsburgh, PA, 15206, USA
| |
Collapse
|
16
|
Villard S, Kidd G. Speech intelligibility and talker gender classification with noise-vocoded and tone-vocoded speech. JASA EXPRESS LETTERS 2021; 1:094401. [PMID: 34590078 PMCID: PMC8456348 DOI: 10.1121/10.0006285] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/02/2021] [Accepted: 08/21/2021] [Indexed: 05/21/2023]
Abstract
Vocoded speech provides less spectral information than natural, unprocessed speech, negatively affecting listener performance on speech intelligibility and talker gender classification tasks. In this study, young normal-hearing participants listened to noise-vocoded and tone-vocoded (i.e., sinewave-vocoded) sentences containing 1, 2, 4, 8, 16, or 32 channels, as well as non-vocoded sentences, and reported the words heard as well as the gender of the talker. Overall, performance was significantly better with tone-vocoded than noise-vocoded speech for both tasks. Within the talker gender classification task, biases in performance were observed for lower numbers of channels, especially when using the noise carrier.
Collapse
Affiliation(s)
- Sarah Villard
- Department of Speech, Language and Hearing Sciences & Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, ,
| | - Gerald Kidd
- Department of Speech, Language and Hearing Sciences & Hearing Research Center, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, ,
| |
Collapse
|
17
|
Fitzhugh MC, LaCroix AN, Rogalsky C. Distinct Contributions of Working Memory and Attentional Control to Sentence Comprehension in Noise in Persons With Stroke. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:3230-3241. [PMID: 34284642 PMCID: PMC8740654 DOI: 10.1044/2021_jslhr-20-00694] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/03/2020] [Revised: 03/26/2021] [Accepted: 04/22/2021] [Indexed: 06/13/2023]
Abstract
Purpose Sentence comprehension deficits are common following a left hemisphere stroke and have primarily been investigated under optimal listening conditions. However, ample work in neurotypical controls indicates that background noise affects sentence comprehension and the cognitive resources it engages. The purpose of this study was to examine how background noise affects sentence comprehension poststroke using both energetic and informational maskers. We further sought to identify whether sentence comprehension in noise abilities are related to poststroke cognitive abilities, specifically working memory and/or attentional control. Method Twenty persons with chronic left hemisphere stroke completed a sentence-picture matching task where they listened to sentences presented in three types of maskers: multispeakers, broadband noise, and silence (control condition). Working memory, attentional control, and hearing thresholds were also assessed. Results A repeated-measures analysis of variance identified participants to have the greatest difficulty with the multispeakers condition, followed by broadband noise and then silence. Regression analyses, after controlling for age and hearing ability, identified working memory as a significant predictor of listening engagement (i.e., mean reaction time) in broadband noise and multispeakers and attentional control as a significant predictor of informational masking effects (computed as a reaction time difference score where broadband noise is subtracted from multispeakers). Conclusions The results from this study indicate that background noise impacts sentence comprehension abilities poststroke and that these difficulties may arise due to deficits in the cognitive resources supporting sentence comprehension and not other factors such as age or hearing. These findings also highlight a relationship between working memory abilities and sentence comprehension in background noise. We further suggest that attentional control abilities contribute to sentence comprehension by supporting the additional demands associated with informational masking. Supplemental Material https://doi.org/10.23641/asha.14984511.
Collapse
Affiliation(s)
- Megan C. Fitzhugh
- Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA
| | | | | |
Collapse
|
18
|
Zanin J, Dhollander T, Rance G, Yu L, Lan L, Wang H, Lou X, Connelly A, Nayagam B, Wang Q. Fiber-Specific Changes in White Matter Microstructure in Individuals With X-Linked Auditory Neuropathy. Ear Hear 2021; 41:1703-1714. [PMID: 33136644 DOI: 10.1097/aud.0000000000000890] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Auditory neuropathy (AN) is the term used to describe a group of hearing disorders, in which the hearing impairment occurs as a result of abnormal auditory nerve function. While our understanding of this condition has advanced significantly over recent years, the ability to determine the site of lesion and the extent of dysfunction in affected individuals remains a challenge. To this end, we investigated potential axonal degeneration in the white matter tracts of the brainstem in individuals with X-linked AN. We hypothesized that individuals with X-linked AN would show focal degeneration within the VIII nerve and/or auditory brainstem tracts, and the degree of degeneration would correlate with the extent of auditory perceptual impairment. DESIGN This was achieved using a higher-order diffusion magnetic resonance imaging (dMRI)-based quantitative measure called apparent fiber density as obtained from a technique called single-shell 3-tissue constrained spherical deconvolution and analyzed with the fixel-based analysis framework. Eleven subjects with genetically confirmed X-linked AN and 11 controls with normal hearing were assessed using behavioral and objective auditory measures. dMRI data were also collected for each participant. RESULTS Fixel-based analysis of the brainstem region showed that subjects with X-linked AN had significantly lower apparent fiber density in the VIII nerve compared with controls, consistent with axonal degeneration in this region. Subsequent analysis of the auditory brainstem tracts specifically showed that degeneration was also significant in these structures overall. The apparent fiber density findings were supported by objective measures of auditory function, such as auditory brainstem responses, electrocochleography, and otoacoustic emissions, which showed VIII nerve activity was severely disrupted in X-linked AN subjects while cochlear sensory hair cell function was relatively unaffected. Moreover, apparent fiber density results were significantly correlated with temporal processing ability (gap detection task) in affected subjects, suggesting that the degree of VIII nerve degeneration may impact the ability to resolve temporal aspects of an acoustic signal. Auditory assessments of sound detection, speech perception, and the processing of binaural cues were also significantly poorer in the X-linked AN group compared with the controls with normal hearing. CONCLUSIONS The results of this study suggest that the dMRI-based measure of apparent fiber density may provide a useful adjunct to existing auditory assessments in the characterization of the site of lesion and extent of dysfunction in individuals with AN. Additionally, the ability to determine the degree of degeneration has the potential to guide rehabilitation strategies in the future.
Collapse
Affiliation(s)
- Julien Zanin
- The HEARing Cooperative Research Centre (HEARing CRC), Melbourne, Victoria, Australia
- Department of Audiology and Speech Pathology, University of Melbourne, Parkville, Australia
| | - Thijs Dhollander
- Developmental Imaging, Murdoch Children's Research Institute, Melbourne, Victoria, Australia
| | - Gary Rance
- The HEARing Cooperative Research Centre (HEARing CRC), Melbourne, Victoria, Australia
- Department of Audiology and Speech Pathology, University of Melbourne, Parkville, Australia
| | - Lan Yu
- Department of Otorhinolaryngology, Head and Neck Surgery, Institute of Otolaryngology, Chinese People's Liberation Army General Hospital 301, Beijing, China
- China National Clinical Research Centre for Otolaryngologic Diseases, Chinese People's Liberation Army General Hospital 301, Beijing, China
| | - Lan Lan
- Department of Otorhinolaryngology, Head and Neck Surgery, Institute of Otolaryngology, Chinese People's Liberation Army General Hospital 301, Beijing, China
- China National Clinical Research Centre for Otolaryngologic Diseases, Chinese People's Liberation Army General Hospital 301, Beijing, China
| | - Hongyang Wang
- Department of Otorhinolaryngology, Head and Neck Surgery, Institute of Otolaryngology, Chinese People's Liberation Army General Hospital 301, Beijing, China
- China National Clinical Research Centre for Otolaryngologic Diseases, Chinese People's Liberation Army General Hospital 301, Beijing, China
| | - Xin Lou
- Department of Radiology, Chinese People's Liberation Army General Hospital 301, Beijing, China
| | - Alan Connelly
- The Florey Institute of Neuroscience and Mental Health, Melbourne, Victoria, Australia
- The Florey Department of Neuroscience and Mental Health, University of Melbourne, Melbourne, Australia
| | - Bryony Nayagam
- Department of Audiology and Speech Pathology, University of Melbourne, Parkville, Australia
- These authors contributed equally to this work
| | - Qiuju Wang
- Department of Otorhinolaryngology, Head and Neck Surgery, Institute of Otolaryngology, Chinese People's Liberation Army General Hospital 301, Beijing, China
- China National Clinical Research Centre for Otolaryngologic Diseases, Chinese People's Liberation Army General Hospital 301, Beijing, China
- Key Lab of Hearing Impairment Science of Ministry of Education, Beijing, China
- These authors contributed equally to this work
| |
Collapse
|
19
|
Wu M, Cañete OM, Schmidt JH, Fereczkowski M, Neher T. Influence of Three Auditory Profiles on Aided Speech Perception in Different Noise Scenarios. Trends Hear 2021; 25:23312165211023709. [PMID: 34184946 PMCID: PMC8246576 DOI: 10.1177/23312165211023709] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2022] Open
Abstract
Hearing aid (HA) users differ greatly in their speech-in-noise (SIN) outcomes. This could be because the degree to which current HA fittings can address individual listening needs differs across users and listening situations. In two earlier studies, an auditory test battery and a data-driven method were developed for classifying HA candidates into four distinct auditory profiles differing in audiometric hearing loss and suprathreshold hearing abilities. This study explored aided SIN outcome for three of these profiles in different noise scenarios. Thirty-one older habitual HA users and six young normal-hearing listeners participated. Two SIN tasks were administered: a speech recognition task and a “just follow conversation” task requiring the participants to self-adjust the target-speech level. Three noise conditions were tested: stationary speech-shaped noise, speech-shaped babble noise, and speech-shaped babble noise with competing dialogues. Each HA user was fitted with three HAs from different manufacturers using their recommended procedures. Real-ear measurements were performed to document the final gain settings. The results showed that HA users with mild hearing deficits performed better than HA users with pronounced hearing deficits on the speech recognition task but not the just follow conversation task. Moreover, participants with pronounced hearing deficits obtained different SIN outcomes with the tested HAs, which appeared to be related to differences in HA gain. Overall, these findings imply that current proprietary fitting strategies are limited in their ability to ensure good SIN outcomes, especially for users with pronounced hearing deficits, for whom the choice of device seems most consequential.
Collapse
Affiliation(s)
- Mengfan Wu
- Institute of Clinical Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark.,Research Unit for ORL-Head & Neck Surgery and Audiology, Odense University Hospital, Odense, Denmark.,University of Southern Denmark, Odense, Denmark
| | - Oscar M Cañete
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
| | - Jesper Hvass Schmidt
- Institute of Clinical Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark.,Research Unit for ORL-Head & Neck Surgery and Audiology, Odense University Hospital, Odense, Denmark.,University of Southern Denmark, Odense, Denmark.,Department of Otolaryngology, Head & Neck Surgery and Audiology, Odense University Hospital, Odense, Denmark.,OPEN, Odense Patient data Explorative Network, Odense University Hospital, Odense, Denmark
| | - Michal Fereczkowski
- Institute of Clinical Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark.,Research Unit for ORL-Head & Neck Surgery and Audiology, Odense University Hospital, Odense, Denmark.,University of Southern Denmark, Odense, Denmark
| | - Tobias Neher
- Institute of Clinical Research, Faculty of Health Sciences, University of Southern Denmark, Odense, Denmark.,Research Unit for ORL-Head & Neck Surgery and Audiology, Odense University Hospital, Odense, Denmark.,University of Southern Denmark, Odense, Denmark
| |
Collapse
|
20
|
Gallun FJ. Impaired Binaural Hearing in Adults: A Selected Review of the Literature. Front Neurosci 2021; 15:610957. [PMID: 33815037 PMCID: PMC8017161 DOI: 10.3389/fnins.2021.610957] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2020] [Accepted: 02/19/2021] [Indexed: 11/17/2022] Open
Abstract
Despite over 100 years of study, there are still many fundamental questions about binaural hearing that remain unanswered, including how impairments of binaural function are related to the mechanisms of binaural hearing. This review focuses on a number of studies that are fundamental to understanding what is known about the effects of peripheral hearing loss, aging, traumatic brain injury, strokes, brain tumors, and multiple sclerosis (MS) on binaural function. The literature reviewed makes clear that while each of these conditions has the potential to impair the binaural system, the specific abilities of a given patient cannot be known without performing multiple behavioral and/or neurophysiological measurements of binaural sensitivity. Future work in this area has the potential to bring awareness of binaural dysfunction to patients and clinicians as well as a deeper understanding of the mechanisms of binaural hearing, but it will require the integration of clinical research with animal and computational modeling approaches.
Collapse
Affiliation(s)
- Frederick J. Gallun
- Oregon Hearing Research Center, Oregon Health and Science University, Portland, OR, United States
| |
Collapse
|
21
|
Mackey CA, McCrate J, MacDonald KS, Feller J, Liberman L, Liberman MC, Hackett TA, Ramachandran R. Correlations between cochlear pathophysiology and behavioral measures of temporal and spatial processing in noise exposed macaques. Hear Res 2021; 401:108156. [PMID: 33373804 PMCID: PMC8487072 DOI: 10.1016/j.heares.2020.108156] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/26/2020] [Revised: 12/10/2020] [Accepted: 12/14/2020] [Indexed: 12/23/2022]
Abstract
Noise-induced hearing loss (NIHL) is known to have significant consequences for temporal, spectral, and spatial resolution. However, much remains to be discovered about their underlying pathophysiology. This report extends the recent development of a nonhuman primate model of NIHL to explore its consequences for hearing in noisy environments, and its correlations with the underlying cochlear pathology. Ten macaques (seven with normal-hearing, three with NIHL) were used in studies of masked tone detection in which the temporal or spatial properties of the masker were varied to assess metrics of temporal and spatial processing. Normal-hearing (NH) macaques showed lower tone detection thresholds for sinusoidally amplitude modulated (SAM) broadband noise maskers relative to unmodulated maskers (modulation masking release, MMR). Tone detection thresholds were lowest at low noise modulation frequencies, and increased as modulation frequency increased, until they matched threshold in unmodulated noise. NH macaques also showed lower tone detection thresholds for spatially separated tone and noise relative to co-localized tone and noise (spatial release from masking, SRM). Noise exposure caused permanent threshold shifts that were verified behaviorally and audiologically. In hearing-impaired (HI) macaques, MMR was reduced at tone frequencies above that of the noise exposure. HI macaques also showed degraded SRM, with no SRM observed across all tested tone frequencies. Deficits in MMR correlated with audiometric threshold changes, outer hair cell loss, and synapse loss, while the differences in SRM did not correlate with audiometric changes, or any measure of cochlear pathophysiology. This difference in anatomical-behavioral correlations suggests that while many behavioral deficits may arise from cochlear pathology, only some are predictable from the frequency place of damage in the cochlea.
Collapse
Affiliation(s)
- Chase A Mackey
- Vanderbilt Neuroscience Graduate Program, Vanderbilt University, Nashville, TN 37212, United States.
| | - Jennifer McCrate
- Interdisciplinary Program in Neuroscience for Undergraduates, Vanderbilt University, Nashville, TN 37240, United States.
| | - Kaitlyn S MacDonald
- Vanderbilt Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| | - Jessica Feller
- Vanderbilt Neuroscience Graduate Program, Vanderbilt University, Nashville, TN 37212, United States.
| | - Leslie Liberman
- Eaton Peabody Laboratories, Massachusetts Eye and Ear Infirmary & Harvard Medical Center, Boston, MA 02114, United States.
| | - M Charles Liberman
- Eaton Peabody Laboratories, Massachusetts Eye and Ear Infirmary & Harvard Medical Center, Boston, MA 02114, United States.
| | - Troy A Hackett
- Vanderbilt Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| | - Ramnarayan Ramachandran
- Vanderbilt Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, United States.
| |
Collapse
|
22
|
Wasiuk PA, Radvansky GA, Greene RL, Calandruccio L. Spoken narrative comprehension for young adult listeners: effects of competing voices and noise. Int J Audiol 2021; 60:711-722. [PMID: 33586551 DOI: 10.1080/14992027.2021.1878397] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
OBJECTIVE To examine the influence of competing voices or noise on the comprehension of spoken narratives for young adults. DESIGN First, an intelligibility assessment of the target narratives was conducted to establish a signal-to-noise ratio ensuring accurate initial speech recognition. Then, narrative comprehension for two target types (fixed and varied target talker) was measured in four listening conditions (quiet, one-talker speech, speech babble, speech-shaped noise). After hearing target narratives in each listening condition, participants completed a visual recognition memory task that assessed the comprehension of the narrative materials at three levels of representation (surface form, propositional, event model). STUDY SAMPLE Seventy adults (18-32 years of age). RESULTS Narrative comprehension results revealed a main effect of listening condition at the event model level, indicating poorer narrative memory of described situations for all noise conditions compared to quiet. Increased positive responses to thematically consistent but situationally "wrong" memory probes drove this effect. No other significant effects were observed. CONCLUSION Despite near-perfect speech recognition, background noise negatively influenced aspects of spoken narrative comprehension and memory. Specifically, noise did not disrupt memory for what was said (surface form and propositional memory), but only memory for what was talked about (event model memory).
Collapse
Affiliation(s)
- Peter A Wasiuk
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH, USA
| | | | - Robert L Greene
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH, USA
| | - Lauren Calandruccio
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH, USA
| |
Collapse
|
23
|
Srinivasan NK, Holtz A, Gallun FJ. Comparing Spatial Release From Masking Using Traditional Methods and Portable Automated Rapid Testing iPad App. Am J Audiol 2020; 29:907-915. [PMID: 33197327 PMCID: PMC8608168 DOI: 10.1044/2020_aja-20-00078] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2020] [Revised: 08/06/2020] [Accepted: 09/02/2020] [Indexed: 11/09/2022] Open
Abstract
Purpose The purpose of this study was to compare speech identification abilities of individuals of various ages and hearing abilities using traditional methods and Portable Automated Rapid Testing (PART) iPad app. Method Speech identification data were collected using three techniques: over headphones using a virtual speaker array, using PART iPad app (UCR Brain Game Center, 2018), and using loudspeaker presentation in a sound-attenuated room. For all three techniques, Coordinate Response Measure sentences were used as the stimuli and "Charlie" was used as the call sign. A progressive tracking procedure was used to estimate the speech identification thresholds for listeners with varying hearing thresholds. The target sentence was always presented at 0° azimuth angle, whereas the maskers were colocated (0°) with the target or symmetrically spatially separated by ±15°, ±30°, or ±45°. Results Data analysis revealed similar speech identification thresholds for the iPad and headphone conditions and slightly poorer thresholds for the loudspeaker array condition across participant groups. This was true for all spatial separations between the target and the maskers. Conclusion Strong correlation between the headphone and iPad data presented in this study indicated that the spatial release from masking module in the PART iPad app can be used as a clinical tool to assess spatial processing ability prior to audiologic evaluation in the clinic and can also be used to make recommendations for and to track progress with aural rehabilitation programs over time.
Collapse
Affiliation(s)
| | - Allison Holtz
- Department of Speech-Language Pathology & Audiology, Towson University, MD
| | - Frederick J. Gallun
- Oregon Health & Science University, Department of Otolaryngology–Head & Neck Surgery, Portland, OR
- Veterans Affairs Rehabilitation Research & Development National Center for Rehabilitative Auditory Research, VA Portland Health Care System, OR
| |
Collapse
|
24
|
Wasiuk PA, Lavandier M, Buss E, Oleson J, Calandruccio L. The effect of fundamental frequency contour similarity on multi-talker listening in older and younger adults. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:3527. [PMID: 33379934 PMCID: PMC7863686 DOI: 10.1121/10.0002661] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Older adults with hearing loss have greater difficulty recognizing target speech in multi-talker environments than young adults with normal hearing, especially when target and masker speech streams are perceptually similar. A difference in fundamental frequency (f0) contour depth is an effective stream segregation cue for young adults with normal hearing. This study examined whether older adults with varying degrees of sensorineural hearing loss are able to utilize differences in target/masker f0 contour depth to improve speech recognition in multi-talker listening. Speech recognition thresholds (SRTs) were measured for speech mixtures composed of target/masker streams with flat, normal, and exaggerated speaking styles, in which f0 contour depth systematically varied. Computational modeling estimated differences in energetic masking across listening conditions. Young adults had lower SRTs than older adults; a result that was partially explained by differences in audibility predicted by the model. However, audibility differences did not explain why young adults experienced a benefit from mismatched target/masker f0 contour depth, while in most conditions, older adults did not. Reduced ability to use segregation cues (differences in target/masker f0 contour depth), and deficits grouping speech with variable f0 contours likely contribute to difficulties experienced by older adults in challenging acoustic environments.
Collapse
Affiliation(s)
- Peter A Wasiuk
- Department of Psychological Sciences, 11635 Euclid Avenue, Case Western Reserve University, Cleveland, Ohio 44106, USA
| | - Mathieu Lavandier
- Univ. Lyon, ENTPE, Laboratoire Génie Civil et Bâtiment, Rue M. Audin, Vaulx-en-Velin Cedex, 69518, France
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina, CB#7070, Chapel Hill, North Carolina 27599, USA
| | - Jacob Oleson
- Department of Biostatistics, N300 CPHB, University of Iowa, 145 North Riverside Drive, Iowa City, Iowa 52242-2007, USA
| | - Lauren Calandruccio
- Department of Psychological Sciences, 11635 Euclid Avenue, Case Western Reserve University, Cleveland, Ohio 44106, USA
| |
Collapse
|
25
|
Villard S, Kidd G. Assessing the benefit of acoustic beamforming for listeners with aphasia using modified psychoacoustic methods. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:2894. [PMID: 33261373 PMCID: PMC8097716 DOI: 10.1121/10.0002454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/02/2020] [Revised: 10/13/2020] [Accepted: 10/14/2020] [Indexed: 06/12/2023]
Abstract
Acoustic beamforming has been shown to improve identification of target speech in noisy listening environments for individuals with sensorineural hearing loss. This study examined whether beamforming would provide a similar benefit for individuals with aphasia (acquired neurological language impairment). The benefit of beamforming was examined for persons with aphasia (PWA) and age- and hearing-matched controls in both a speech masking condition and a speech-shaped, speech-modulated noise masking condition. Performance was measured when natural spatial cues were provided, as well as when the target speech level was enhanced via a single-channel beamformer. Because typical psychoacoustic methods may present substantial experimental confounds for PWA, clinically guided modifications of experimental procedures were determined individually for each PWA participant. Results indicated that the beamformer provided a significant overall benefit to listeners. On an individual level, both PWA and controls who exhibited poorer performance on the speech masking condition with spatial cues benefited from the beamformer, while those who achieved better performance with spatial cues did not. All participants benefited from the beamformer in the noise masking condition. The findings suggest that a spatially tuned hearing aid may be beneficial for older listeners with relatively mild hearing loss who have difficulty taking advantage of spatial cues.
Collapse
Affiliation(s)
- Sarah Villard
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| | - Gerald Kidd
- Department of Speech, Language, and Hearing Sciences, Boston University, 635 Commonwealth Avenue, Boston, Massachusetts 02215, USA
| |
Collapse
|
26
|
Fitzhugh MC, Schaefer SY, Baxter LC, Rogalsky C. Cognitive and neural predictors of speech comprehension in noisy backgrounds in older adults. LANGUAGE, COGNITION AND NEUROSCIENCE 2020; 36:269-287. [PMID: 34250179 PMCID: PMC8261331 DOI: 10.1080/23273798.2020.1828946] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/06/2019] [Accepted: 09/18/2020] [Indexed: 06/13/2023]
Abstract
Older adults often experience difficulties comprehending speech in noisy backgrounds, which hearing loss does not fully explain. It remains unknown how cognitive abilities, brain networks, and age-related hearing loss may uniquely contribute to speech in noise comprehension at the sentence level. In 31 older adults, using cognitive measures and resting-state fMRI, we investigated the cognitive and neural predictors of speech comprehension with energetic (broadband noise) and informational masking (multi-speakers) effects. Better hearing thresholds and greater working memory abilities were associated with better speech comprehension with energetic masking. Conversely, faster processing speed and stronger functional connectivity between frontoparietal and language networks were associated with better speech comprehension with informational masking. Our findings highlight the importance of the frontoparietal network in older adults' ability to comprehend speech in multi-speaker backgrounds, and that hearing loss and working memory in older adults contributes to speech comprehension abilities related to energetic, but not informational masking.
Collapse
Affiliation(s)
- Megan C. Fitzhugh
- Stevens Neuroimaging and Informatics Institute, University of Southern California, Los Angeles, CA
- College of Health Solutions, Arizona State University, Tempe, AZ
| | - Sydney Y. Schaefer
- School of Biological and Health Systems Engineering, Arizona State University, Tempe, AZ
| | | | | |
Collapse
|
27
|
Song J, Martin L, Iverson P. Auditory neural tracking and lexical processing of speech in noise: Masker type, spatial location, and language experience. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 148:253. [PMID: 32752786 DOI: 10.1121/10.0001477] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/11/2019] [Accepted: 06/09/2020] [Indexed: 06/11/2023]
Abstract
The present study investigated how single-talker and babble maskers affect auditory and lexical processing during native (L1) and non-native (L2) speech recognition. Electroencephalogram (EEG) recordings were made while L1 and L2 (Korean) English speakers listened to sentences in the presence of single-talker and babble maskers that were colocated or spatially separated from the target. The predictability of the sentences was manipulated to measure lexical-semantic processing (N400), and selective auditory processing of the target was assessed using neural tracking measures. The results demonstrate that intelligible single-talker maskers cause listeners to attend more to the semantic content of the targets (i.e., greater context-related N400 changes) than when targets are in babble, and that listeners track the acoustics of the target less accurately with single-talker maskers. L1 and L2 listeners both modulated their processing in this way, although L2 listeners had more difficulty with the materials overall (i.e., lower behavioral accuracy, less context-related N400 variation, more listening effort). The results demonstrate that auditory and lexical processing can be simultaneously assessed within a naturalistic speech listening task, and listeners can adjust lexical processing to more strongly track the meaning of a sentence in order to help ignore competing lexical content.
Collapse
Affiliation(s)
- Jieun Song
- Department of Speech, Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London, WC1N 1PF, United Kingdom
| | - Luke Martin
- Department of Speech, Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London, WC1N 1PF, United Kingdom
| | - Paul Iverson
- Department of Speech, Hearing and Phonetic Sciences, University College London, Chandler House, 2 Wakefield Street, London, WC1N 1PF, United Kingdom
| |
Collapse
|
28
|
Adel Ghahraman M, Ashrafi M, Mohammadkhani G, Jalaie S. Effects of aging on spatial hearing. Aging Clin Exp Res 2020; 32:733-739. [PMID: 31203530 DOI: 10.1007/s40520-019-01233-3] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2018] [Accepted: 05/28/2019] [Indexed: 11/28/2022]
Abstract
BACKGROUND Aging has several effects on auditory processing with the most important effect known as speech perception impairment in noise. AIMS The aim of the present study was to investigate the effects of aging on spatial hearing using quick speech in noise (QSIN) and binaural masking level difference (BMLD) tests and speech, spatial, and qualities of hearing scale (SSQ) questionnaire. METHODS The study was carried out on 34 elderly people, aged 60-75 years, with normal peripheral hearing and 34 young participants, aged 18-25 years. Using SSQ questionnaire and QSIN and BMLD tests, the spatial auditory processing ability was compared between the two groups. RESULTS Comparison of mean scores using independent t test showed that there was a significant difference in the mean scores of QSIN, BMLD tests and SSQ questionnaire between the two groups (p < 0.001). Sex was not found to have any effect on the results (p > 0.05). DISCUSSION Structural and neurochemical changes that occur in different parts of the central nervous system by aging affect various aspects of spatial auditory processing, such as localization, the precedence effect, and speech perception in noise. CONCLUSIONS Lower scores of older adults with normal hearing in SSQ questionnaire and behavioral tests, compared with younger participants, may be considered as their weak performance in spatial auditory processing. The results of the present study reconfirm the effects of aging on spatial auditory processing, such as localization and speech perception in noise.
Collapse
Affiliation(s)
- Mansoureh Adel Ghahraman
- Department of Audiology, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| | - Majid Ashrafi
- Department of Audiology, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| | - Ghassem Mohammadkhani
- Department of Audiology, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran.
| | - Shohreh Jalaie
- Biostatistics, School of Rehabilitation, Tehran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
29
|
Masked Sentence Recognition in Children, Young Adults, and Older Adults: Age-Dependent Effects of Semantic Context and Masker Type. Ear Hear 2020; 40:1117-1126. [PMID: 30601213 DOI: 10.1097/aud.0000000000000692] [Citation(s) in RCA: 29] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Masked speech recognition in normal-hearing listeners depends in part on masker type and semantic context of the target. Children and older adults are more susceptible to masking than young adults, particularly when the masker is speech. Semantic context has been shown to facilitate noise-masked sentence recognition in all age groups, but it is not known whether age affects a listener's ability to use context with a speech masker. The purpose of the present study was to evaluate the effect of masker type and semantic context of the target as a function of listener age. DESIGN Listeners were children (5 to 16 years), young adults (19 to 30 years), and older adults (67 to 81 years), all with normal or near-normal hearing. Maskers were either speech-shaped noise or two-talker speech, and targets were either semantically correct (high context) sentences or semantically anomalous (low context) sentences. RESULTS As predicted, speech reception thresholds were lower for young adults than either children or older adults. Age effects were larger for the two-talker masker than the speech-shaped noise masker, and the effect of masker type was larger in children than older adults. Performance tended to be better for targets with high than low semantic context, but this benefit depended on age group and masker type. In contrast to adults, children benefitted less from context in the two-talker speech masker than the speech-shaped noise masker. Context effects were small compared with differences across age and masker type. CONCLUSIONS Different effects of masker type and target context are observed at different points across the lifespan. While the two-talker masker is particularly challenging for children and older adults, the speech masker may limit the use of semantic context in children but not adults.
Collapse
|
30
|
Ahrens A, Marschall M, Dau T. The effect of spatial energy spread on sound image size and speech intelligibility. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:1368. [PMID: 32237851 DOI: 10.1121/10.0000747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/04/2019] [Accepted: 01/30/2020] [Indexed: 06/11/2023]
Abstract
This study explored the relationship between perceived sound image size and speech intelligibility for sound sources reproduced over loudspeakers. Sources with varying degrees of spatial energy spread were generated using ambisonics processing. Young normal-hearing listeners estimated sound image size as well as performed two spatial release from masking (SRM) tasks with two symmetrically arranged interfering talkers. Either the target-to-masker ratio or the separation angle was varied adaptively. Results showed that the sound image size did not change systematically with the energy spread. However, a larger energy spread did result in a decreased SRM. Furthermore, the listeners needed a greater angular separation angle between the target and the interfering sources for sources with a larger energy spread. Further analysis revealed that the method employed to vary the energy spread did not lead to systematic changes in the interaural cross correlations. Future experiments with competing talkers using ambisonics or similar methods may consider the resulting energy spread in relation to the minimum separation angle between sound sources in order to avoid degradations in speech intelligibility.
Collapse
Affiliation(s)
- Axel Ahrens
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Building 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| | - Marton Marschall
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Building 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, Building 352, Ørsteds Plads, 2800 Kongens Lyngby, Denmark
| |
Collapse
|
31
|
Domingo Y, Holmes E, Macpherson E, Johnsrude IS. Using spatial release from masking to estimate the magnitude of the familiar-voice intelligibility benefit. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:3487. [PMID: 31795686 DOI: 10.1121/1.5133628] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/18/2019] [Accepted: 10/23/2019] [Indexed: 06/10/2023]
Abstract
The ability to segregate simultaneous speech streams is crucial for successful communication. Recent studies have demonstrated that participants can report 10%-20% more words spoken by naturally familiar (e.g., friends or spouses) than unfamiliar talkers in two-voice mixtures. This benefit is commensurate with one of the largest benefits to speech intelligibility currently known-that which is gained by spatially separating two talkers. However, because of differences in the methods of these previous studies, the relative benefits of spatial separation and voice familiarity are unclear. Here, the familiar-voice benefit and spatial release from masking are directly compared, and it is examined if and how these two cues interact with one another. Talkers were recorded while speaking sentences from a published closed-set "matrix" task, and then listeners were presented with three different sentences played simultaneously. Each target sentence was played at 0° azimuth, and two masker sentences were symmetrically separated about the target. On average, participants reported 10%-30% more words correctly when the target sentence was spoken in a familiar than unfamiliar voice (collapsed over spatial separation conditions); it was found that participants gain a similar benefit from a familiar target as when an unfamiliar voice is separated from two symmetrical maskers by approximately 15° azimuth.
Collapse
Affiliation(s)
- Ysabel Domingo
- Brain and Mind Institute, University of Western Ontario, London, Ontario, Canada
| | - Emma Holmes
- Brain and Mind Institute, University of Western Ontario, London, Ontario, Canada
| | - Ewan Macpherson
- School of Communication Sciences and Disorders, University of Western Ontario, London, Ontario, Canada
| | - Ingrid S Johnsrude
- Brain and Mind Institute, University of Western Ontario, London, Ontario, Canada
| |
Collapse
|
32
|
Muñoz RV, Aspöck L, Fels J. Spatial Release From Masking Under Different Reverberant Conditions in Young and Elderly Subjects: Effect of Moving or Stationary Maskers at Circular and Radial Conditions. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:3582-3595. [PMID: 31525113 DOI: 10.1044/2019_jslhr-h-19-0092] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/22/2023]
Abstract
Purpose Normal-hearing and hard-of-hearing listeners suffer from reduced speech intelligibility in noisy and reverberant environments. Although daily listening environments are in constant motion, most researchers have only studied speech-in-noise perception for stationary masker locations. The aim of this study was to investigate the spatial release from masking (SRM) of circularly and radially moving maskers under different room acoustic conditions for young and elderly subjects. Method Twelve young subjects with normal hearing and 12 elderly subjects with normal hearing or mild hearing loss were tested. Several different room acoustic conditions were simulated and reproduced via headphones using binaural synthesis. The target speech stream consisted of German digit triplets, and masker stream consisted of quasistationary noise with matched long-term averaged speech spectra. During the experiment, the position of the masker was changed to be in different stationary positions, or varied continuously. In the latter case, it was moved either on a circular trajectory spanning a 90° azimuth angle or on a radial trajectory linearly increasing the distance to the receiver from 0.5 m to 1.8 m. Absorption characteristics of the virtual room's surfaces were changed, recreating an anechoic room, a treated room with mean reverberation times (RT60) = 0.48 s, and an untreated room with mean RT60 = 1.26 s. Results For the circular condition, a significant difference was found between moving and stationary maskers, F(4, 44) = 20.91, p < .001, with a bigger SRM for stationary maskers than moving masker conditions. Also, both age groups displayed a significant decrease in SRM over the reverberation conditions: F(2, 22) = 12.24, p < .001. For the radial condition, both age groups showed a significant decrease in SRM over the reverberation conditions, F(2, 22) = 13.62, p < .001, as well as the moving and stationary masker conditions, F(8, 88) = 29.23, p < .001. In general, the SRM of a moving masker decreased when the reverberation increased, especially for elderly subjects. Conclusions A radially moving masker led to improved SRM in an anechoic environment for both age groups, whereas a circularly moving masker caused degraded SRM, especially for elderly subjects in the highly reverberant environment. Supplemental Material https://doi.org/10.23641/asha.9795371.
Collapse
Affiliation(s)
- Rhoddy Viveros Muñoz
- Teaching and Research Area of Medical Acoustics, Institute of Technical Acoustics, RWTH Aachen University, Germany
| | - Lukas Aspöck
- Chair and Institute of Technical Acoustics, RWTH Aachen University, Germany
| | - Janina Fels
- Teaching and Research Area of Medical Acoustics, Institute of Technical Acoustics, RWTH Aachen University, Germany
| |
Collapse
|
33
|
Abstract
OBJECTIVE The clinical evaluation of hearing loss, using a pure-tone audiogram, is not adequate to assess the functional hearing capabilities (or handicap) of a patient, especially the speech-in-noise communication difficulties. The primary objective of this study was to measure the effect of elevated hearing thresholds on the recognition performance in various functional speech-in-noise tests that cover acoustic scenes of different complexities and to identify the subset of tests that (a) were sensitive to individual differences in hearing thresholds and (b) provide complementary information to the audiogram. A secondary goal was to compare the performance on this test battery with the self-assessed performance level of functional hearing abilities. DESIGN In this study, speech-in-noise performance of normal-hearing listeners and listeners with hearing loss (audiometric configuration ranging from near-normal hearing to moderate-severe hearing loss) was measured on a battery of 12 different tests designed to evaluate speech recognition in a variety of speech and masker conditions, and listening tasks. The listening conditions were designed to measure the ability to localize and monitor multiple speakers or to take advantage of masker modulation, spatial separation between the target and the masker, and a restricted vocabulary. RESULTS Listeners with hearing loss had significantly worse performance than the normal-hearing control group when speech was presented in the presence of a multitalker babble or in the presence of a single competing talker. In particular, the ability to take advantage of modulation benefit and spatial release from masking was significantly affected even with a mild audiometric loss. Elevated thresholds did not have a significant effect on the performance in the spatial awareness task. A composite score of all 12 tests was considered as a global metric of the overall speech-in-noise performance. Perceived hearing difficulties of subjects were better correlated with the composite score than with the performance on a standardized clinical speech-in-noise test. Regression analysis showed that scores from a subset of these tests, which could potentially take less than 10 min to administer, when combined with the better-ear pure-tone average and the subject's age, accounted for as much as 93.2% of the variance in the composite score. CONCLUSIONS A test that measures speech recognition in the presence of a spatially separated competing talker would be useful in measuring suprathreshold speech-in-noise deficits that cannot be readily predicted from standard audiometric evaluation. Including such a test can likely reduce the gap between patient complaints and their clinical evaluation.
Collapse
|
34
|
Zobel BH, Wagner A, Sanders LD, Başkent D. Spatial release from informational masking declines with age: Evidence from a detection task in a virtual separation paradigm. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:548. [PMID: 31370625 DOI: 10.1121/1.5118240] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/19/2018] [Accepted: 06/28/2019] [Indexed: 06/10/2023]
Abstract
Declines in spatial release from informational masking may contribute to the speech-processing difficulties that older adults often experience within complex listening environments. The present study sought to answer two fundamental questions: (1) Does spatial release from informational masking decline with age and, if so, (2) does age predict this decline independently of age-typical hearing loss? Younger (18-34 years) and older (60-80 years) adults with age-typical hearing completed a yes/no target-detection task with low-pass filtered noise-vocoded speech designed to reduce non-spatial segregation cues and control for hearing loss. Participants detected a target voice among two-talker masking babble while a virtual spatial separation paradigm [Freyman, Helfer, McCall, and Clifton, J. Acoust. Soc. Am. 106(6), 3578-3588 (1999)] was used to isolate informational masking release. The younger and older adults both exhibited spatial release from informational masking, but masking release was reduced among the older adults. Furthermore, age predicted this decline controlling for hearing loss, while there was no indication that hearing loss played a role. These findings provide evidence that declines specific to aging limit spatial release from informational masking under challenging listening conditions.
Collapse
Affiliation(s)
- Benjamin H Zobel
- Department of Psychological and Brain Sciences, University of Massachusetts, Amherst, Massachusetts 01003, USA
| | - Anita Wagner
- Department of Otorhinolaryngology-Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, the Netherlands
| | - Lisa D Sanders
- Department of Psychological and Brain Sciences, University of Massachusetts, Amherst, Massachusetts 01003, USA
| | - Deniz Başkent
- Department of Otorhinolaryngology-Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, the Netherlands
| |
Collapse
|
35
|
Redundant Information Is Sometimes More Beneficial Than Spatial Information to Understand Speech in Noise. Ear Hear 2019; 40:545-554. [DOI: 10.1097/aud.0000000000000660] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
36
|
Deroche MLD, Gracco VL. Segregation of voices with single or double fundamental frequencies. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:847. [PMID: 30823786 DOI: 10.1121/1.5090107] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2018] [Accepted: 01/23/2019] [Indexed: 06/09/2023]
Abstract
In cocktail-party situations, listeners can use the fundamental frequency (F0) of a voice to segregate it from competitors, but other cues in speech could help, such as co-modulation of envelopes across frequency or more complex cues related to the semantic/syntactic content of the utterances. For simplicity, this (non-pitch) form of grouping is referred to as "articulatory." By creating a new type of speech with two steady F0s, it was examined how these two forms of segregation compete: articulatory grouping would bind the partials of a double-F0 source together, whereas harmonic segregation would tend to split them in two subsets. In experiment 1, maskers were two same-male sentences. Speech reception thresholds were high in this task (vicinity of 0 dB), and harmonic segregation behaved as though double-F0 stimuli were two independent sources. This was not the case in experiment 2, where maskers were speech-shaped complexes (buzzes). First, double-F0 targets were immune to the masking of a single-F0 buzz matching one of the two target F0s. Second, double-F0 buzzes were particularly effective at masking a single-F0 target matching one of the two buzz F0s. As a conclusion, the strength of F0-segregation appears to depend on whether the masker is speech or not.
Collapse
Affiliation(s)
- Mickael L D Deroche
- Centre for Research on Brain, Language and Music, McGill University, 3640 rue de la Montagne, Montreal, H3G 2A8, Canada
| | - Vincent L Gracco
- Haskins Laboratories, 300 George Street, New Haven, Connecticut 06511, USA
| |
Collapse
|
37
|
Kidd G, Mason CR, Best V, Roverud E, Swaminathan J, Jennings T, Clayton K, Steven Colburn H. Determining the energetic and informational components of speech-on-speech masking in listeners with sensorineural hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:440. [PMID: 30710924 PMCID: PMC6347574 DOI: 10.1121/1.5087555] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/20/2018] [Revised: 11/19/2018] [Accepted: 12/18/2018] [Indexed: 05/20/2023]
Abstract
The ability to identify the words spoken by one talker masked by two or four competing talkers was tested in young-adult listeners with sensorineural hearing loss (SNHL). In a reference/baseline condition, masking speech was colocated with target speech, target and masker talkers were female, and the masker was intelligible. Three comparison conditions included replacing female masker talkers with males, time-reversal of masker speech, and spatial separation of sources. All three variables produced significant release from masking. To emulate energetic masking (EM), stimuli were subjected to ideal time-frequency segregation retaining only the time-frequency units where target energy exceeded masker energy. Subjects were then tested with these resynthesized "glimpsed stimuli." For either two or four maskers, thresholds only varied about 3 dB across conditions suggesting that EM was roughly equal. Compared to normal-hearing listeners from an earlier study [Kidd, Mason, Swaminathan, Roverud, Clayton, and Best, J. Acoust. Soc. Am. 140, 132-144 (2016)], SNHL listeners demonstrated both greater energetic and informational masking as well as higher glimpsed thresholds. Individual differences were correlated across masking release conditions suggesting that listeners could be categorized according to their general ability to solve the task. Overall, both peripheral and central factors appear to contribute to the higher thresholds for SNHL listeners.
Collapse
Affiliation(s)
- Gerald Kidd
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Christine R Mason
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Virginia Best
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Elin Roverud
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Jayaganesh Swaminathan
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Todd Jennings
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Kameron Clayton
- Department of Speech, Language and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - H Steven Colburn
- Department of Biomedical Engineering, Boston University, Boston, Massachusetts 02215, USA
| |
Collapse
|
38
|
Auditory cortex responses to interaural time differences in the envelope of low-frequency sound, recorded with MEG in young and older listeners. Hear Res 2018; 370:22-39. [PMID: 30265860 DOI: 10.1016/j.heares.2018.09.001] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/16/2018] [Revised: 08/31/2018] [Accepted: 09/03/2018] [Indexed: 11/21/2022]
Abstract
Interaural time and intensity differences (ITD and IID) are important cues in binaural hearing and allow for sound localization, improving speech understanding in noise and reverberation, and integrating sound sources in the auditory scene. Whereas previous research showed that the upper-frequency limit for ITD detection in the fine structure of sound declines in aging, the processing of envelope ITD in low-frequency amplitude modulated (AM) sound and the related brain responses are less understood. This study investigated the cortical processing of envelope ITD and compared the results with previous findings about the fine-structure ITD. In two experiments, participants listened to 40-Hz AM tones containing sudden changes in the envelope ITD. Multiple MEG responses were analyzed, including the auditory evoked N1 responses, elicited both by sound onsets and ITD changes, and 40-Hz responses, elicited by the AM. The first experiment with healthy young adults revealed a substantial decline in the magnitudes of the ITD change N1 response, and the 40-Hz phase resets at higher carrier frequencies, suggesting a similar frequency characteristic as observed for fine structure ITD. The amplitude of the 40-Hz ASSR declined only gradually with increasing carrier frequency, and it was excluded as a confounding factor in the decline in the ITD response. Larger responses to outward ITD changes than inward changes, here first reported for envelope ITD, were another characteristics that were similar to fine-structure ITD. A second experiment with groups of young and older listeners examined the effects of aging and concurrent noise on the cortical envelope ITD responses. One important research question was, whether binaural cues are accessible in noise. Behavioural tests showed an age-related hearing loss in the older group and decreased performance in envelope ITD detection and speech-in-noise (SIN) understanding. Binaural hearing and SIN performance were correlated with one other, but not with hearing loss. The frequency limit for envelope ITD was reduced in older listeners similarly as previously found for fine structure ITD, and older listeners were more susceptible to concurrent multi-talker noise. The similarities between responses to envelope ITD and to fine structure ITD suggest that a common cortical code exists for the envelope and fine structure ITD. The dependency on the carrier frequency suggests that envelope ITD processing at the subcortical level requires stimulus phase locking, which might be reduced in aging.
Collapse
|
39
|
Dong Y, Raif KE, Determan SC, Gai Y. Decoding spatial attention with EEG and virtual acoustic space. Physiol Rep 2018; 5:5/22/e13512. [PMID: 29180483 PMCID: PMC5704085 DOI: 10.14814/phy2.13512] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2017] [Accepted: 10/30/2017] [Indexed: 11/24/2022] Open
Abstract
Decoding spatial attention based on brain signals has wide applications in brain–computer interface (BCI). Previous BCI systems mostly relied on visual patterns or auditory stimulation (e.g., loudspeakers) to evoke synchronous brain signals. There would be difficulties to cover a large range of spatial locations with such a stimulation protocol. The present study explored the possibility of using virtual acoustic space and a visual‐auditory matching paradigm to overcome this issue. The technique has the flexibility of generating sound stimulation from virtually any spatial location. Brain signals of eight human subjects were obtained with a 32‐channel Electroencephalogram (EEG). Two amplitude‐modulated noise or speech sentences carrying distinct spatial information were presented concurrently. Each sound source was tagged with a unique modulation phase so that the phase of the recorded EEG signals indicated the sound being attended to. The phase‐tagged sound was further filtered with head‐related transfer functions to create the sense of virtual space. Subjects were required to pay attention to the sound source that best matched the location of a visual target. For all the subjects, the phase of a single sound could be accurately reflected over the majority of electrodes based on EEG responses of 90 s or less. The electrodes providing significant decoding performance on auditory attention were fewer and may require longer EEG responses. The reliability and efficiency of decoding with a single electrode varied with subjects. Overall, the virtual acoustic space protocol has the potential of being used in practical BCI systems.
Collapse
Affiliation(s)
- Yue Dong
- Biomedical Engineering Department, Parks College of Engineering, Aviation and Technology Saint Louis University, St Louis, Missouri
| | - Kaan E Raif
- Biomedical Engineering Department, Parks College of Engineering, Aviation and Technology Saint Louis University, St Louis, Missouri
| | - Sarah C Determan
- Biomedical Engineering Department, Parks College of Engineering, Aviation and Technology Saint Louis University, St Louis, Missouri
| | - Yan Gai
- Biomedical Engineering Department, Parks College of Engineering, Aviation and Technology Saint Louis University, St Louis, Missouri
| |
Collapse
|
40
|
Rana B, Buchholz JM. Effect of audibility on better-ear glimpsing as a function of frequency in normal-hearing and hearing-impaired listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 143:2195. [PMID: 29716302 DOI: 10.1121/1.5031007] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Better-ear glimpsing (BEG) is an auditory phenomenon that helps understanding speech in noise by utilizing interaural level differences (ILDs). The benefit provided by BEG is limited in hearing-impaired (HI) listeners by reduced audibility at high frequencies. Rana and Buchholz [(2016). J. Acoust. Soc. Am. 140(2), 1192-1205] have shown that artificially enhancing ILDs at low and mid frequencies can help HI listeners understanding speech in noise, but the achieved benefit is smaller than in normal-hearing (NH) listeners. To understand how far this difference is explained by differences in audibility, audibility was carefully controlled here in ten NH and ten HI listeners and speech reception thresholds (SRTs) in noise were measured in a spatially separated and co-located condition as a function of frequency and sensation level. Maskers were realized by noise-vocoded speech and signals were spatialized using artificially generated broadband ILDs. The spatial benefit provided by BEG and SRTs improved consistently with increasing sensation level, but was limited in the HI listeners by loudness discomfort. Further, the HI listeners performed similar to NH listeners when differences in audibility were compensated. The results help to understand the hearing aid gain that is required to maximize the spatial benefit provided by ILDs as a function of frequency.
Collapse
Affiliation(s)
- Baljeet Rana
- National Acoustic Laboratories, 16 University Avenue, Macquarie University, Sydney, New South Wales 2109, Australia
| | - Jörg M Buchholz
- National Acoustic Laboratories, 16 University Avenue, Macquarie University, Sydney, New South Wales 2109, Australia
| |
Collapse
|
41
|
Davis TJ, Gifford RH. Spatial Release From Masking in Adults With Bilateral Cochlear Implants: Effects of Distracter Azimuth and Microphone Location. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:752-761. [PMID: 29450488 PMCID: PMC5963045 DOI: 10.1044/2017_jslhr-h-16-0441] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2016] [Revised: 08/20/2017] [Accepted: 10/04/2017] [Indexed: 06/01/2023]
Abstract
PURPOSE The primary purpose of this study was to derive spatial release from masking (SRM) performance-azimuth functions for bilateral cochlear implant (CI) users to provide a thorough description of SRM as a function of target/distracter spatial configuration. The secondary purpose of this study was to investigate the effect of the microphone location for SRM in a within-subject study design. METHOD Speech recognition was measured in 12 adults with bilateral CIs for 11 spatial separations ranging from -90° to +90° in 20° steps using an adaptive block design. Five of the 12 participants were tested with both the behind-the-ear microphones and a T-mic configuration to further investigate the effect of mic location on SRM. RESULTS SRM can be significantly affected by the hemifield origin of the distracter stimulus-particularly for listeners with interaural asymmetry in speech understanding. The greatest SRM was observed with a distracter positioned 50° away from the target. There was no effect of mic location on SRM for the current experimental design. CONCLUSION Our results demonstrate that the traditional assessment of SRM with a distracter positioned at 90° azimuth may underestimate maximum performance for individuals with bilateral CIs.
Collapse
Affiliation(s)
- Timothy J. Davis
- Department of Hearing and Speech Sciences, Vanderbilt University, Nashville, TN
| | - René H. Gifford
- Department of Hearing and Speech Sciences, Vanderbilt University, Nashville, TN
| |
Collapse
|
42
|
Corbin NE, Buss E, Leibold LJ. Spatial Release From Masking in Children: Effects of Simulated Unilateral Hearing Loss. Ear Hear 2018; 38:223-235. [PMID: 27787392 PMCID: PMC5321780 DOI: 10.1097/aud.0000000000000376] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES The purpose of this study was twofold: (1) to determine the effect of an acute simulated unilateral hearing loss on children's spatial release from masking in two-talker speech and speech-shaped noise, and (2) to develop a procedure to be used in future studies that will assess spatial release from masking in children who have permanent unilateral hearing loss. There were three main predictions. First, spatial release from masking was expected to be larger in two-talker speech than in speech-shaped noise. Second, simulated unilateral hearing loss was expected to worsen performance in all listening conditions, but particularly in the spatially separated two-talker speech masker. Third, spatial release from masking was expected to be smaller for children than for adults in the two-talker masker. DESIGN Participants were 12 children (8.7 to 10.9 years) and 11 adults (18.5 to 30.4 years) with normal bilateral hearing. Thresholds for 50%-correct recognition of Bamford-Kowal-Bench sentences were measured adaptively in continuous two-talker speech or speech-shaped noise. Target sentences were always presented from a loudspeaker at 0° azimuth. The masker stimulus was either co-located with the target or spatially separated to +90° or -90° azimuth. Spatial release from masking was quantified as the difference between thresholds obtained when the target and masker were co-located and thresholds obtained when the masker was presented from +90° or -90° azimuth. Testing was completed both with and without a moderate simulated unilateral hearing loss, created with a foam earplug and supra-aural earmuff. A repeated-measures design was used to compare performance between children and adults, and performance in the no-plug and simulated-unilateral-hearing-loss conditions. RESULTS All listeners benefited from spatial separation of target and masker stimuli on the azimuth plane in the no-plug listening conditions; this benefit was larger in two-talker speech than in speech-shaped noise. In the simulated-unilateral-hearing-loss conditions, a positive spatial release from masking was observed only when the masker was presented ipsilateral to the simulated unilateral hearing loss. In the speech-shaped noise masker, spatial release from masking in the no-plug condition was similar to that obtained when the masker was presented ipsilateral to the simulated unilateral hearing loss. In contrast, in the two-talker speech masker, spatial release from masking in the no-plug condition was much larger than that obtained when the masker was presented ipsilateral to the simulated unilateral hearing loss. When either masker was presented contralateral to the simulated unilateral hearing loss, spatial release from masking was negative. This pattern of results was observed for both children and adults, although children performed more poorly overall. CONCLUSIONS Children and adults with normal bilateral hearing experience greater spatial release from masking for a two-talker speech than a speech-shaped noise masker. Testing in a two-talker speech masker revealed listening difficulties in the presence of disrupted binaural input that were not observed in a speech-shaped noise masker. This procedure offers promise for the assessment of spatial release from masking in children with permanent unilateral hearing loss.
Collapse
Affiliation(s)
- Nicole E. Corbin
- Department of Allied Health Sciences, Division of Speech and Hearing Sciences, University of North Carolina at Chapel Hill, School of Medicine, Chapel Hill, NC, USA
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina at Chapel Hill, School of Medicine, Chapel Hill, NC, USA
| | | |
Collapse
|
43
|
Jakien KM, Kampel SD, Gordon SY, Gallun FJ. The Benefits of Increased Sensation Level and Bandwidth for Spatial Release From Masking. Ear Hear 2018; 38:e13-e21. [PMID: 27556520 PMCID: PMC5161636 DOI: 10.1097/aud.0000000000000352] [Citation(s) in RCA: 25] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2014] [Accepted: 06/03/2016] [Indexed: 11/26/2022]
Abstract
OBJECTIVE Spatial release from masking (SRM) can increase speech intelligibility in complex listening environments. The goal of the present study was to document how speech-in-speech stimuli could be best processed to encourage optimum SRM for listeners who represent a range of ages and amounts of hearing loss. We examined the effects of equating stimulus audibility among listeners, presenting stimuli at uniform sensation levels (SLs), and filtering stimuli at two separate bandwidths. DESIGN Seventy-one participants completed two speech intelligibility experiments (36 listeners in experiment 1; all 71 in experiment 2) in which a target phrase from the coordinate response measure (CRM) and two masking phrases from the CRM were presented simultaneously via earphones using a virtual spatial array, such that the target sentence was always at 0 degree azimuth angle and the maskers were either colocated or positioned at ±45 degrees. Experiments 1 and 2 examined the impacts of SL, age, and hearing loss on SRM. Experiment 2 also assessed the effects of stimulus bandwidth on SRM. RESULTS Overall, listeners' ability to achieve SRM improved with increased SL. Younger listeners with less hearing loss achieved more SRM than older or hearing-impaired listeners. It was hypothesized that SL and bandwidth would result in dissociable effects on SRM. However, acoustical analysis revealed that effective audible bandwidth, defined as the highest frequency at which the stimulus was audible at both ears, was the best predictor of performance. Thus, increasing SL seemed to improve SRM by increasing the effective bandwidth rather than increasing the level of already audible components. CONCLUSIONS Performance for all listeners, regardless of age or hearing loss, improved with an increase in overall SL and/or bandwidth, but the improvement was small relative to the benefits of spatial separation.
Collapse
Affiliation(s)
- Kasey M. Jakien
- Otolaryngology/Head & Neck Surgery, Oregon Health & Science University, Portland, Oregon, USA; and Department of Veterans Affairs, Portland VA Medical Center, National Center for Rehabilitative Auditory Research, Portland, Oregon, USA
| | - Sean D. Kampel
- Otolaryngology/Head & Neck Surgery, Oregon Health & Science University, Portland, Oregon, USA; and Department of Veterans Affairs, Portland VA Medical Center, National Center for Rehabilitative Auditory Research, Portland, Oregon, USA
| | - Samuel Y. Gordon
- Otolaryngology/Head & Neck Surgery, Oregon Health & Science University, Portland, Oregon, USA; and Department of Veterans Affairs, Portland VA Medical Center, National Center for Rehabilitative Auditory Research, Portland, Oregon, USA
| | - Frederick J. Gallun
- Otolaryngology/Head & Neck Surgery, Oregon Health & Science University, Portland, Oregon, USA; and Department of Veterans Affairs, Portland VA Medical Center, National Center for Rehabilitative Auditory Research, Portland, Oregon, USA
| |
Collapse
|
44
|
Graydon K, Van Dun B, Tomlin D, Dowell R, Rance G. Remediation of spatial processing disorder (SPD). Int J Audiol 2018; 57:376-384. [DOI: 10.1080/14992027.2018.1431403] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Affiliation(s)
- Kelley Graydon
- The HEARing Cooperative Research Centre, Carlton, Victoria, Australia,
- Department of Audiology and Speech Pathology, The University of Melbourne, Carlton, Victoria, Australia, and
| | - Bram Van Dun
- The HEARing Cooperative Research Centre, Carlton, Victoria, Australia,
- National Acoustic Laboratories, Macquarie Park, New South Wales, Australia
| | - Dani Tomlin
- The HEARing Cooperative Research Centre, Carlton, Victoria, Australia,
- Department of Audiology and Speech Pathology, The University of Melbourne, Carlton, Victoria, Australia, and
| | - Richard Dowell
- Department of Audiology and Speech Pathology, The University of Melbourne, Carlton, Victoria, Australia, and
| | - Gary Rance
- Department of Audiology and Speech Pathology, The University of Melbourne, Carlton, Victoria, Australia, and
| |
Collapse
|
45
|
Courtois G, Lissek H, Estoppey P, Oesch Y, Gigandet X. Effects of Binaural Spatialization in Wireless Microphone Systems for Hearing Aids on Normal-Hearing and Hearing-Impaired Listeners. Trends Hear 2018; 22:2331216517753548. [PMID: 29457537 PMCID: PMC5821302 DOI: 10.1177/2331216517753548] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2017] [Revised: 12/07/2017] [Accepted: 12/12/2017] [Indexed: 11/17/2022] Open
Abstract
Little is known about the perception of artificial spatial hearing by hearing-impaired subjects. The purpose of this study was to investigate how listeners with hearing disorders perceived the effect of a spatialization feature designed for wireless microphone systems. Forty listeners took part in the experiments. They were arranged in four groups: normal-hearing, moderate, severe, and profound hearing loss. Their performance in terms of speech understanding and speaker localization was assessed with diotic and binaural stimuli. The results of the speech intelligibility experiment revealed that the subjects presenting a moderate or severe hearing impairment better understood speech with the spatialization feature. Thus, it was demonstrated that the conventional diotic binaural summation operated by current wireless systems can be transformed to reproduce the spatial cues required to localize the speaker, without any loss of intelligibility. The speaker localization experiment showed that a majority of the hearing-impaired listeners had similar performance with natural and artificial spatial hearing, contrary to the normal-hearing listeners. This suggests that certain subjects with hearing impairment preserve their localization abilities with approximated generic head-related transfer functions in the frontal horizontal plane.
Collapse
Affiliation(s)
- Gilles Courtois
- Swiss Federal Institute of Technology, Signal Processing Laboratory, Lausanne, Switzerland
| | - Hervé Lissek
- Swiss Federal Institute of Technology, Signal Processing Laboratory, Lausanne, Switzerland
| | | | - Yves Oesch
- Phonak Communications AG, Murten, Switzerland
| | | |
Collapse
|
46
|
Yang X, Jiang M, Zhao Y. Effects of Noise on English Listening Comprehension among Chinese College Students with Different Learning Styles. Front Psychol 2017; 8:1764. [PMID: 29085317 PMCID: PMC5650695 DOI: 10.3389/fpsyg.2017.01764] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2017] [Accepted: 09/25/2017] [Indexed: 11/13/2022] Open
Abstract
This study was intended to determine whether the effects of noise on English listening comprehension would vary among Chinese college students with different learning styles. A total of 89 participants with different learning styles measured using Kolb’s (1985) Learning Style Inventory finished English listening comprehension tests in quiet and in white noise, Chinese two-talker babble, and English two-talker babble respectively. The results showed that the participants in general had significantly poorer performance in the two babble conditions than in quiet and white noise. However, the participants with assimilative and divergent learning styles performed relatively better in Chinese babble, and exhibited stable performance across the three noisy conditions, while the participants with convergent and accommodative learning styles had more impaired performance in both Chinese babble and English babble than in white noise. Moreover, of Kolb’s four learning modes, reflective observation had a facilitative effect on listening performance in Chinese babble and English babble. These findings suggest that differences in learning style might lead to differential performance in foreign language listening comprehension in noise.
Collapse
Affiliation(s)
- Xiaohu Yang
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, Shanghai, China
| | - Meng Jiang
- Language & Brain Research Center, Sichuan International Studies University, Chongqing, China
| | - Yong Zhao
- Department of Translation and Interpreting, School of Foreign Languages, Shanghai Jiao Tong University, Shanghai, China
| |
Collapse
|
47
|
Helfer KS, Merchant GR, Wasiuk PA. Age-Related Changes in Objective and Subjective Speech Perception in Complex Listening Environments. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2017; 60:3009-3018. [PMID: 29049601 PMCID: PMC5945070 DOI: 10.1044/2017_jslhr-h-17-0030] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2017] [Revised: 05/03/2017] [Accepted: 05/03/2017] [Indexed: 06/07/2023]
Abstract
PURPOSE A frequent complaint by older adults is difficulty communicating in challenging acoustic environments. The purpose of this work was to review and summarize information about how speech perception in complex listening situations changes across the adult age range. METHOD This article provides a review of age-related changes in speech understanding in complex listening environments and summarizes results from several studies conducted in our laboratory. RESULTS Both degree of high frequency hearing loss and cognitive test performance limit individuals' ability to understand speech in difficult listening situations as they age. The performance of middle-aged adults is similar to that of younger adults in the presence of noise maskers, but they experience substantially more difficulty when the masker is 1 or 2 competing speech messages. For the most part, middle-aged participants in studies conducted in our laboratory reported as much self-perceived hearing problems as did older adult participants. CONCLUSIONS Research supports the multifactorial nature of listening in real-world environments. Current audiologic assessment practices are often insufficient to identify the true speech understanding struggles that individuals experience in these situations. This points to the importance of giving weight to patients' self-reported difficulties. PRESENTATION VIDEO http://cred.pubs.asha.org/article.aspx?articleid=2601619.
Collapse
Affiliation(s)
- Karen S. Helfer
- Department of Communication Disorders, University of Massachusetts Amherst
| | | | - Peter A. Wasiuk
- Department of Communication Disorders, University of Massachusetts Amherst
| |
Collapse
|
48
|
Bissmeyer SRS, Goldsworthy RL. Adaptive spatial filtering improves speech reception in noise while preserving binaural cues. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:1441. [PMID: 28964069 PMCID: PMC8267853 DOI: 10.1121/1.5002691] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2015] [Revised: 08/08/2017] [Accepted: 08/28/2017] [Indexed: 06/02/2023]
Abstract
Hearing loss greatly reduces an individual's ability to comprehend speech in the presence of background noise. Over the past decades, numerous signal-processing algorithms have been developed to improve speech reception in these situations for cochlear implant and hearing aid users. One challenge is to reduce background noise while not introducing interaural distortion that would degrade binaural hearing. The present study evaluates a noise reduction algorithm, referred to as binaural Fennec, that was designed to improve speech reception in background noise while preserving binaural cues. Speech reception thresholds were measured for normal-hearing listeners in a simulated environment with target speech generated in front of the listener and background noise originating 90° to the right of the listener. Lateralization thresholds were also measured in the presence of background noise. These measures were conducted in anechoic and reverberant environments. Results indicate that the algorithm improved speech reception thresholds, even in highly reverberant environments. Results indicate that the algorithm also improved lateralization thresholds for the anechoic environment while not affecting lateralization thresholds for the reverberant environments. These results provide clear evidence that this algorithm can improve speech reception in background noise while preserving binaural cues used to lateralize sound.
Collapse
Affiliation(s)
- Susan R S Bissmeyer
- Caruso Department of Otolaryngology, Caruso Center for Childhood Communication, University of Southern California, 806 West Adams Boulevard, Los Angeles, California 90007, USA
| | - Raymond L Goldsworthy
- Caruso Department of Otolaryngology, Caruso Center for Childhood Communication, University of Southern California, 806 West Adams Boulevard, Los Angeles, California 90007, USA
| |
Collapse
|
49
|
Role of Binaural Temporal Fine Structure and Envelope Cues in Cocktail-Party Listening. J Neurosci 2017; 36:8250-7. [PMID: 27488643 DOI: 10.1523/jneurosci.4421-15.2016] [Citation(s) in RCA: 32] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2015] [Accepted: 06/19/2016] [Indexed: 11/21/2022] Open
Abstract
UNLABELLED While conversing in a crowded social setting, a listener is often required to follow a target speech signal amid multiple competing speech signals (the so-called "cocktail party" problem). In such situations, separation of the target speech signal in azimuth from the interfering masker signals can lead to an improvement in target intelligibility, an effect known as spatial release from masking (SRM). This study assessed the contributions of two stimulus properties that vary with separation of sound sources, binaural envelope (ENV) and temporal fine structure (TFS), to SRM in normal-hearing (NH) human listeners. Target speech was presented from the front and speech maskers were either colocated with or symmetrically separated from the target in azimuth. The target and maskers were presented either as natural speech or as "noise-vocoded" speech in which the intelligibility was conveyed only by the speech ENVs from several frequency bands; the speech TFS within each band was replaced with noise carriers. The experiments were designed to preserve the spatial cues in the speech ENVs while retaining/eliminating them from the TFS. This was achieved by using the same/different noise carriers in the two ears. A phenomenological auditory-nerve model was used to verify that the interaural correlations in TFS differed across conditions, whereas the ENVs retained a high degree of correlation, as intended. Overall, the results from this study revealed that binaural TFS cues, especially for frequency regions below 1500 Hz, are critical for achieving SRM in NH listeners. Potential implications for studying SRM in hearing-impaired listeners are discussed. SIGNIFICANCE STATEMENT Acoustic signals received by the auditory system pass first through an array of physiologically based band-pass filters. Conceptually, at the output of each filter, there are two principal forms of temporal information: slowly varying fluctuations in the envelope (ENV) and rapidly varying fluctuations in the temporal fine structure (TFS). The importance of these two types of information in everyday listening (e.g., conversing in a noisy social situation; the "cocktail-party" problem) has not been established. This study assessed the contributions of binaural ENV and TFS cues for understanding speech in multiple-talker situations. Results suggest that, whereas the ENV cues are important for speech intelligibility, binaural TFS cues are critical for perceptually segregating the different talkers and thus for solving the cocktail party problem.
Collapse
|
50
|
Rocchi F, Dylla ME, Bohlen PA, Ramachandran R. Spatial and temporal disparity in signals and maskers affects signal detection in non-human primates. Hear Res 2017; 344:1-12. [PMID: 27770624 PMCID: PMC5239734 DOI: 10.1016/j.heares.2016.10.013] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/22/2016] [Revised: 10/10/2016] [Accepted: 10/15/2016] [Indexed: 12/25/2022]
Abstract
Detection thresholds for auditory stimuli (signals) increase in the presence of maskers. Natural environments contain maskers/distractors that can have a wide range of spatiotemporal properties relative to the signal. While these parameters have been well explored psychophysically in humans, they have not been well explored in animal models, and their neuronal underpinnings are not well understood. As a precursor to the neuronal measurements, we report the effects of systematically varying the spatial and temporal relationship between signals and noise in macaque monkeys (Macaca mulatta and Macaca radiata). Macaques detected tones masked by noise in a Go/No-Go task in which the spatiotemporal relationships between the tone and noise were systematically varied. Masked thresholds were higher when the masker was continuous or gated on and off simultaneously with the signal, and lower when the continuous masker was turned off during the signal. A burst of noise caused higher masked thresholds if it completely temporally overlapped with the signal, whereas partial overlap resulted in lower thresholds. Noise durations needed to be at least 100 ms before significant masking could be observed. Thresholds for short duration tones were significantly higher when the onsets of signal and masker coincided compared to when the signal was presented during the steady state portion of the noise (overshoot). When signal and masker were separated in space, masked signal detection thresholds decreased relative to when the masker and signal were co-located (spatial release from masking). Masking release was larger for azimuthal separations than for elevation separations. These results in macaques are similar to those observed in humans, suggesting that the specific spatiotemporal relationship between signal and masker determine threshold in natural environments for macaques in a manner similar to humans. These results form the basis for future investigations of neuronal correlates and mechanisms of masking.
Collapse
Affiliation(s)
- Francesca Rocchi
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37212, USA
| | - Margit E Dylla
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37212, USA
| | - Peter A Bohlen
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37212, USA
| | - Ramnarayan Ramachandran
- Department of Hearing and Speech Sciences, Vanderbilt University Medical Center, Nashville, TN 37212, USA.
| |
Collapse
|