1
|
Fumero MJ, Marrufo-Pérez MI, Eustaquio-Martín A, Lopez-Poveda EA. Factors that can affect divided speech intelligibility. Hear Res 2024; 441:108917. [PMID: 38061268 DOI: 10.1016/j.heares.2023.108917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Revised: 11/10/2023] [Accepted: 11/16/2023] [Indexed: 12/27/2023]
Abstract
Previous studies have shown that in challenging listening situations, people find it hard to equally divide their attention between two simultaneous talkers and tend to favor one talker over the other. The aim here was to investigate whether talker onset/offset, sex and location determine the favored talker. Fifteen people with normal hearing were asked to recognize as many words as possible from two sentences uttered by two talkers located at 45° and +45° azimuth, respectively. The sentences were from the same corpus, were time-centered and had equal sound level. In Conditions 1 and 2, the talkers had different sexes (male at +45°), sentence duration was not controlled for, and sentences were presented at 65 and 35 dB SPL, respectively. Listeners favored the male over the female talker, even more so at 35 dB SPL (62 % vs 43 % word recognition, respectively) than at 65 dB SPL (74 % vs 64 %, respectively). The greater asymmetry in intelligibility at the lower level supports that divided listening is harder and more 'asymmetric' in challenging acoustic scenarios. Listeners continued to favor the male talker when the experiment was repeated with sentences of equal average duration for the two talkers (Condition 3). This suggests that the earlier onset or later offset of male sentences (52 ms on average) was not the reason for the asymmetric intelligibility in Conditions 1 or 2. When the location of the talkers was switched (Condition 4) or the two talkers were the same woman (Condition 5), listeners continued to favor the talker to their right albeit non-significantly. Altogether, results confirm that in hard divided listening situations, listeners tend to favor the talker to their right. This preference is not affected by talker onset/offset delays less than 52 ms on average. Instead, the preference seems to be modulated by the voice characteristics of the talkers.
Collapse
Affiliation(s)
- Milagros J Fumero
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Miriam I Marrufo-Pérez
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Almudena Eustaquio-Martín
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain
| | - Enrique A Lopez-Poveda
- Instituto de Neurociencias de Castilla y León, Universidad de Salamanca, Salamanca 37007, Spain; Instituto de Investigación Biomédica de Salamanca, Universidad de Salamanca, Salamanca 37007, Spain; Departamento de Cirugía, Facultad de Medicina, Universidad de Salamanca, Salamanca 37007, Spain.
| |
Collapse
|
2
|
Shim YJ, Jung WH, Billig AJ, Sedley W, Song JJ. Hippocampal atrophy is associated with hearing loss in cognitively normal adults. Front Neurosci 2023; 17:1276883. [PMID: 37942139 PMCID: PMC10628109 DOI: 10.3389/fnins.2023.1276883] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2023] [Accepted: 10/11/2023] [Indexed: 11/10/2023] Open
Abstract
Objectives A growing body of evidence suggests that age-related hearing loss (HL) is associated with morphological changes of the cerebral cortex, but the results have been drawn from a small amount of data in most studies. The aim of this study is to investigate the correlation between HL and gray matter volume (GMV) in a large number of subjects, strictly controlling for an extensive set of possible biases. Methods Medical records of 576 subjects who underwent pure tone audiometry, brain magnetic resonance imaging (MRI), and the Korean Mini-Mental State Exam (K-MMSE) were reviewed. Among them, subjects with normal cognitive function and free of central nervous system disorders or coronary artery disease were included. Outliers were excluded after a sample homogeneity check. In the end, 405 subjects were enrolled. Pure tone hearing thresholds were determined at 0.5, 1, 2, and 4 kHz in the better ear. Enrolled subjects were divided into 3 groups according to pure tone average: normal hearing (NH), mild HL (MHL), and moderate-to-severe HL (MSHL) groups. Using voxel-based morphometry, we evaluated GMV changes that may be associated with HL. Sex, age, total intracranial volume, type of MRI scanner, education level, K-MMSE score, smoking status, and presence of hypertension, diabetes mellitus and dyslipidemia were used as covariates. Results A statistically significant negative correlation between the hearing thresholds and GMV of the hippocampus was elucidated. Additionally, in group comparisons, the left hippocampal GMV of the MSHL group was significantly smaller than that of the NH and MHL groups. Conclusion Based on the negative correlation between hearing thresholds and hippocampal GMV in cognitively normal old adults, the current study indicates that peripheral deafferentation could be a potential contributing factor to hippocampal atrophy.
Collapse
Affiliation(s)
- Ye Ji Shim
- Department of Otorhinolaryngology-Head and Neck Surgery, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, Republic of Korea
- Sensory Organ Research Institute, Seoul National University Medical Research Center, Seoul, Republic of Korea
| | - Wi Hoon Jung
- Department of Psychology, Gachon University, Seongnam, Republic of Korea
| | | | - William Sedley
- Translational and Clinical Research Institute, Newcastle University Medical School, Newcastle upon Tyne, United Kingdom
| | - Jae-Jin Song
- Sensory Organ Research Institute, Seoul National University Medical Research Center, Seoul, Republic of Korea
- Department of Otorhinolaryngology-Head and Neck Surgery, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| |
Collapse
|
3
|
Neuschwander P, Schmitt R, Jagoda L, Kurthen I, Giroud N, Meyer M. Different neuroanatomical correlates for temporal and spectral supra-threshold auditory tasks and speech in noise recognition in older adults with hearing impairment. Eur J Neurosci 2023; 57:981-1002. [PMID: 36683390 DOI: 10.1111/ejn.15922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2021] [Revised: 08/20/2022] [Accepted: 01/06/2023] [Indexed: 01/24/2023]
Abstract
Varying degrees of pure-tone hearing loss in older adults are differentially associated with cortical volume (CV) and thickness (CT) within and outside of the auditory pathway. This study addressed the question to what degree supra-threshold auditory performance (i.e., temporal compression and frequency selectivity) as well as speech in noise (SiN) recognition are associated with neurostructural correlates in a sample of 59 healthy older adults with mild to moderate pure-tone hearing loss. Using surface-based morphometry on T1-weighted MRI images, CT, CV, and surface area (CSA) of several regions-of-interest were obtained. The results showed distinct neurostructural patterns for the different tasks in terms of involved regions as well as morphometric parameters. While pure-tone averages (PTAs) positively correlated with CT in a right hemisphere superior temporal sulcus and gyrus cluster, supra-threshold auditory perception additionally extended significantly to CV and CT in left and right superior temporal clusters including Heschl's gyrus and sulcus, the planum polare and temporale. For SiN recognition, we found significant correlations with an auditory-related CT cluster and furthermore with language-related areas in the prefrontal cortex. Taken together, our results show that different auditory abilities are differently associated with cortical morphology in older adults with hearing impairment. Still, a common pattern is that greater PTAs and poorer supra-threshold auditory performance as well as poorer SiN recognition are all related to cortical thinning and volume loss but not to changes in CSA. These results support the hypothesis that mostly CT undergoes alterations in the context of auditory decline, while CSA remains stable.
Collapse
Affiliation(s)
- Pia Neuschwander
- Division of Neuropsychology, Department of Psychology, University of Zurich, Zurich, Switzerland
| | - Raffael Schmitt
- Neuroscience of Speech & Hearing, Department of Computational Linguistics, University of Zurich, Zurich, Switzerland
| | - Laura Jagoda
- Division of Neuropsychology, Department of Psychology, University of Zurich, Zurich, Switzerland
| | - Ira Kurthen
- Developmental Psychology: Infancy and Childhood, Department of Psychology, University of Zurich, Zurich, Switzerland
| | - Nathalie Giroud
- Neuroscience of Speech & Hearing, Department of Computational Linguistics, University of Zurich, Zurich, Switzerland
| | - Martin Meyer
- Evolutionary Neuroscience of Language, Department of Comparative Language Science, University of Zurich, Zurich, Switzerland.,Center for the Interdisciplinary Study of Language Evolution (ISLE), University of Zurich, Zurich, Switzerland.,Cognitive Psychology Unit, Alpen-Adria University of Klagenfurt, Klagenfurt, Austria
| |
Collapse
|
4
|
Cochlear Implant Facilitates the Use of Talker Sex and Spatial Cues to Segregate Competing Speech in Unilaterally Deaf Listeners. Ear Hear 2023; 44:77-91. [PMID: 35733275 DOI: 10.1097/aud.0000000000001254] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
OBJECTIVES Talker sex and spatial cues can facilitate segregation of competing speech. However, the spectrotemporal degradation associated with cochlear implants (CIs) can limit the benefit of talker sex and spatial cues. Acoustic hearing in the nonimplanted ear can improve access to talker sex cues in CI users. However, it's unclear whether the CI can improve segregation of competing speech when maskers are symmetrically placed around the target (i.e., when spatial cues are available), compared with acoustic hearing alone. The aim of this study was to investigate whether a CI can improve segregation of competing speech by individuals with unilateral hearing loss. DESIGN Speech recognition thresholds (SRTs) for competing speech were measured in 16 normal-hearing (NH) adults and 16 unilaterally deaf CI users. All participants were native speakers of Mandarin Chinese. CI users were divided into two groups according to thresholds in the nonimplanted ear: (1) single-sided deaf (SSD); pure-tone thresholds <25 dB HL at all audiometric frequencies, and (2) Asymmetric hearing loss (AHL; one or more thresholds > 25 dB HL). SRTs were measured for target sentences produced by a male talker in the presence of two masker talkers (different male or female talkers). The target sentence was always presented via loudspeaker directly in front of the listener (0°), and the maskers were either colocated with the target (0°) or spatially separated from the target at ±90°. Three segregation cue conditions were tested to measure masking release (MR) relative to the baseline condition: (1) Talker sex, (2) Spatial, and (3) Talker sex + Spatial. For CI users, SRTs were measured with the CI on or off. RESULTS Binaural MR was significantly better for the NH group than for the AHL or SSD groups ( P < 0.001 in all cases). For the NH group, mean MR was largest with the Talker sex + spatial cues (18.8 dB) and smallest for the Talker sex cues (10.7 dB). In contrast, mean MR for the SSD group was largest with the Talker sex + spatial cues (14.7 dB), and smallest with the Spatial cues (4.8 dB). For the AHL group, mean MR was largest with the Talker sex + spatial cues (7.8 dB) and smallest with the Talker sex (4.8 dB) and the Spatial cues (4.8 dB). MR was significantly better with the CI on than off for both the AHL ( P = 0.014) and SSD groups ( P < 0.001). Across all unilaterally deaf CI users, monaural (acoustic ear alone) and binaural MR were significantly correlated with unaided pure-tone average thresholds in the nonimplanted ear for the Talker sex and Talker sex + spatial conditions ( P < 0.001 in both cases) but not for the Spatial condition. CONCLUSION Although the CI benefitted unilaterally deaf listeners' segregation of competing speech, MR was much poorer than that observed in NH listeners. Different from previous findings with steady noise maskers, the CI benefit for segregation of competing speech from a different talker sex was greater in the SSD group than in the AHL group.
Collapse
|
5
|
Pinto D, Kaufman M, Brown A, Zion Golumbic E. An ecological investigation of the capacity to follow simultaneous speech and preferential detection of ones’ own name. Cereb Cortex 2022; 33:5361-5374. [PMID: 36331339 DOI: 10.1093/cercor/bhac424] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2022] [Revised: 09/11/2022] [Accepted: 09/12/2022] [Indexed: 11/06/2022] Open
Abstract
Abstract
Many situations require focusing attention on one speaker, while monitoring the environment for potentially important information. Some have proposed that dividing attention among 2 speakers involves behavioral trade-offs, due to limited cognitive resources. However the severity of these trade-offs, particularly under ecologically-valid circumstances, is not well understood. We investigated the capacity to process simultaneous speech using a dual-task paradigm simulating task-demands and stimuli encountered in real-life. Participants listened to conversational narratives (Narrative Stream) and monitored a stream of announcements (Barista Stream), to detect when their order was called. We measured participants’ performance, neural activity, and skin conductance as they engaged in this dual-task. Participants achieved extremely high dual-task accuracy, with no apparent behavioral trade-offs. Moreover, robust neural and physiological responses were observed for target-stimuli in the Barista Stream, alongside significant neural speech-tracking of the Narrative Stream. These results suggest that humans have substantial capacity to process simultaneous speech and do not suffer from insufficient processing resources, at least for this highly ecological task-combination and level of perceptual load. Results also confirmed the ecological validity of the advantage for detecting ones’ own name at the behavioral, neural, and physiological level, highlighting the contribution of personal relevance when processing simultaneous speech.
Collapse
Affiliation(s)
- Danna Pinto
- The Gonda Multidisciplinary Center for Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel
| | - Maya Kaufman
- The Gonda Multidisciplinary Center for Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel
| | - Adi Brown
- The Gonda Multidisciplinary Center for Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel
| | - Elana Zion Golumbic
- The Gonda Multidisciplinary Center for Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel
| |
Collapse
|
6
|
Divided listening in the free field becomes asymmetric when acoustic cues are limited. Hear Res 2022; 416:108444. [DOI: 10.1016/j.heares.2022.108444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/17/2021] [Revised: 01/10/2022] [Accepted: 01/16/2022] [Indexed: 11/23/2022]
|
7
|
Wisniewski MG, Zakrzewski AC, Bell DR, Wheeler M. EEG power spectral dynamics associated with listening in adverse conditions. Psychophysiology 2021; 58:e13877. [PMID: 34161612 DOI: 10.1111/psyp.13877] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2020] [Revised: 05/15/2021] [Accepted: 05/17/2021] [Indexed: 01/08/2023]
Abstract
Adverse listening conditions increase the demand on cognitive resources needed for speech comprehension. In an exploratory study, we aimed to identify independent power spectral features in the EEG useful for studying the cognitive processes involved in this effortful listening. Listeners performed the coordinate response measure task with a single-talker masker at a 0-dB signal-to-noise ratio. Sounds were left unfiltered or degraded with low-pass filtering. Independent component analysis (ICA) was used to identify independent components (ICs) in the EEG data, the power spectral dynamics of which were then analyzed. Frontal midline theta, left frontal, right frontal, left mu, right mu, left temporal, parietal, left occipital, central occipital, and right occipital clusters of ICs were identified. All IC clusters showed some significant listening-related changes in their power spectrum. This included sustained theta enhancements, gamma enhancements, alpha enhancements, alpha suppression, beta enhancements, and mu rhythm suppression. Several of these effects were absent or negligible using traditional channel analyses. Comparison of filtered to unfiltered speech revealed a stronger alpha suppression in the parietal and central occipital clusters of ICs for the filtered speech condition. This not only replicates recent findings showing greater alpha suppression as listening difficulty increases but also suggests that such alpha-band effects can stem from multiple cortical sources. We lay out the advantages of the ICA approach over the restrictive analyses that have been used as of late in the study of listening effort. We also make suggestions for moving into hypothesis-driven studies regarding the power spectral features that were revealed.
Collapse
Affiliation(s)
- Matthew G Wisniewski
- Department of Psychological Sciences, Kansas State University, Manhattan, KS, USA
| | | | - Destiny R Bell
- Department of Psychological Sciences, Kansas State University, Manhattan, KS, USA
| | - Michelle Wheeler
- Department of Psychological Sciences, Kansas State University, Manhattan, KS, USA
| |
Collapse
|
8
|
Alickovic E, Ng EHN, Fiedler L, Santurette S, Innes-Brown H, Graversen C. Effects of Hearing Aid Noise Reduction on Early and Late Cortical Representations of Competing Talkers in Noise. Front Neurosci 2021; 15:636060. [PMID: 33841081 PMCID: PMC8032942 DOI: 10.3389/fnins.2021.636060] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Accepted: 02/26/2021] [Indexed: 11/13/2022] Open
Abstract
OBJECTIVES Previous research using non-invasive (magnetoencephalography, MEG) and invasive (electrocorticography, ECoG) neural recordings has demonstrated the progressive and hierarchical representation and processing of complex multi-talker auditory scenes in the auditory cortex. Early responses (<85 ms) in primary-like areas appear to represent the individual talkers with almost equal fidelity and are independent of attention in normal-hearing (NH) listeners. However, late responses (>85 ms) in higher-order non-primary areas selectively represent the attended talker with significantly higher fidelity than unattended talkers in NH and hearing-impaired (HI) listeners. Motivated by these findings, the objective of this study was to investigate the effect of a noise reduction scheme (NR) in a commercial hearing aid (HA) on the representation of complex multi-talker auditory scenes in distinct hierarchical stages of the auditory cortex by using high-density electroencephalography (EEG). DESIGN We addressed this issue by investigating early (<85 ms) and late (>85 ms) EEG responses recorded in 34 HI subjects fitted with HAs. The HA noise reduction (NR) was either on or off while the participants listened to a complex auditory scene. Participants were instructed to attend to one of two simultaneous talkers in the foreground while multi-talker babble noise played in the background (+3 dB SNR). After each trial, a two-choice question about the content of the attended speech was presented. RESULTS Using a stimulus reconstruction approach, our results suggest that the attention-related enhancement of neural representations of target and masker talkers located in the foreground, as well as suppression of the background noise in distinct hierarchical stages is significantly affected by the NR scheme. We found that the NR scheme contributed to the enhancement of the foreground and of the entire acoustic scene in the early responses, and that this enhancement was driven by better representation of the target speech. We found that the target talker in HI listeners was selectively represented in late responses. We found that use of the NR scheme resulted in enhanced representations of the target and masker speech in the foreground and a suppressed representation of the noise in the background in late responses. We found a significant effect of EEG time window on the strengths of the cortical representation of the target and masker. CONCLUSION Together, our analyses of the early and late responses obtained from HI listeners support the existing view of hierarchical processing in the auditory cortex. Our findings demonstrate the benefits of a NR scheme on the representation of complex multi-talker auditory scenes in different areas of the auditory cortex in HI listeners.
Collapse
Affiliation(s)
- Emina Alickovic
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- Department of Electrical Engineering, Linkoping University, Linkoping, Sweden
| | - Elaine Hoi Ning Ng
- Centre for Applied Audiology Research, Oticon A/S, Smørum, Denmark
- Department of Behavioral Sciences and Learning, Linkoping University, Linkoping, Sweden
| | - Lorenz Fiedler
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
| | - Sébastien Santurette
- Centre for Applied Audiology Research, Oticon A/S, Smørum, Denmark
- Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
| | | | | |
Collapse
|
9
|
Venezia JH, Leek MR, Lindeman MP. Suprathreshold Differences in Competing Speech Perception in Older Listeners With Normal and Impaired Hearing. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2020; 63:2141-2161. [PMID: 32603618 DOI: 10.1044/2020_jslhr-19-00324] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Purpose Age-related declines in auditory temporal processing and cognition make older listeners vulnerable to interference from competing speech. This vulnerability may be increased in older listeners with sensorineural hearing loss due to additional effects of spectral distortion and accelerated cognitive decline. The goal of this study was to uncover differences between older hearing-impaired (OHI) listeners and older normal-hearing (ONH) listeners in the perceptual encoding of competing speech signals. Method Age-matched groups of 10 OHI and 10 ONH listeners performed the coordinate response measure task with a synthetic female target talker and a male competing talker at a target-to-masker ratio of +3 dB. Individualized gain was provided to OHI listeners. Each listener completed 50 baseline and 800 "bubbles" trials in which randomly selected segments of the speech modulation power spectrum (MPS) were retained on each trial while the remainder was filtered out. Average performance was fixed at 50% correct by adapting the number of segments retained. Multinomial regression was used to estimate weights showing the regions of the MPS associated with performance (a "classification image" or CImg). Results The CImg weights were significantly different between the groups in two MPS regions: a region encoding the shared phonetic content of the two talkers and a region encoding the competing (male) talker's voice. The OHI listeners demonstrated poorer encoding of the phonetic content and increased vulnerability to interference from the competing talker. Individual differences in CImg weights explained over 75% of the variance in baseline performance in the OHI listeners, whereas differences in high-frequency pure-tone thresholds explained only 10%. Conclusion Suprathreshold deficits in the encoding of low- to mid-frequency (~5-10 Hz) temporal modulations-which may reflect poorer "dip listening"-and auditory grouping at a perceptual and/or cognitive level are responsible for the relatively poor performance of OHI versus ONH listeners on a different-gender competing speech task. Supplemental Material https://doi.org/10.23641/asha.12568472.
Collapse
Affiliation(s)
- Jonathan H Venezia
- VA Loma Linda Healthcare System, CA
- Department of Otolaryngology-Head and Neck Surgery, School of Medicine, Loma Linda University, CA
| | - Marjorie R Leek
- VA Loma Linda Healthcare System, CA
- Department of Otolaryngology-Head and Neck Surgery, School of Medicine, Loma Linda University, CA
| | | |
Collapse
|
10
|
Paying attention to speech: The role of working memory capacity and professional experience. Atten Percept Psychophys 2020; 82:3594-3605. [PMID: 32676806 DOI: 10.3758/s13414-020-02091-2] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Managing attention in multispeaker environments is a challenging feat that is critical for human performance. However, why some people are better than others in allocating attention appropriately remains highly unknown. Here, we investigated the contribution of two factors-working memory capacity (WMC) and professional experience-to performance on two different types of attention task: selective attention to one speaker and distributed attention among multiple concurrent speakers. We compared performance across three groups: individuals with low (n = 20) and high (n = 25) WMC, and aircraft pilots (n = 24), whose profession poses extremely high demands for both selective and distributed attention to speech. Results suggests that selective attention is highly effective, with good performance maintained under increasingly adverse conditions, whereas performance decreases substantially with the requirement to distribute attention among a larger number of speakers. Importantly, both types of attention benefit from higher WMC, suggesting reliance on some common capacity-limited resources. However, only selective attention was further improved in the pilots, pointing to its flexible and trainable nature, whereas distributed attention seems to suffer from more fixed and severe processing bottlenecks.
Collapse
|
11
|
Chiu F, Rakusen LL, Mattys SL. Cognitive load elevates discrimination thresholds of duration, intensity, and f 0 for a synthesized vowel. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 146:1077. [PMID: 31472597 DOI: 10.1121/1.5120404] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/03/2019] [Accepted: 07/15/2019] [Indexed: 06/10/2023]
Abstract
Dual-tasking negatively impacts on speech perception by raising cognitive load (CL). Previous research has shown that CL increases reliance on lexical knowledge and decreases reliance on phonetic detail. Less is known about the effect of CL on the perception of acoustic dimensions below the phonetic level. This study tested the effect of CL on the ability to discriminate differences in duration, intensity, and fundamental frequency of a synthesized vowel. A psychophysical adaptive procedure was used to obtain just noticeable differences (JNDs) on each dimension under load and no load. Load was imposed by N-back tasks at two levels of difficulty (one-back, two-back) and under two types of load (images, nonwords). Compared to a control condition with no CL, all N-back conditions increased JNDs across the three dimensions. JNDs were also higher under two-back than one-back load. Nonword load was marginally more detrimental than image load for intensity and fundamental frequency discrimination. Overall, the decreased auditory acuity demonstrates that the effect of CL on the listening experience can be traced to distortions in the perception of core auditory dimensions.
Collapse
Affiliation(s)
- Faith Chiu
- Department of Psychology, University of York, York, YO10 5DD, United Kingdom
| | - Lyndon L Rakusen
- Department of Psychology, University of York, York, YO10 5DD, United Kingdom
| | - Sven L Mattys
- Department of Psychology, University of York, York, YO10 5DD, United Kingdom
| |
Collapse
|
12
|
Parthasarathy A, Bartlett EL, Kujawa SG. Age-related Changes in Neural Coding of Envelope Cues: Peripheral Declines and Central Compensation. Neuroscience 2019; 407:21-31. [DOI: 10.1016/j.neuroscience.2018.12.007] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2018] [Revised: 11/30/2018] [Accepted: 12/03/2018] [Indexed: 12/22/2022]
|
13
|
Meister H, Rählmann S, Walger M. Low background noise increases cognitive load in older adults listening to competing speech. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:EL417. [PMID: 30522293 DOI: 10.1121/1.5078953] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/25/2018] [Accepted: 10/28/2018] [Indexed: 06/09/2023]
Abstract
This letter describes a dual-task paradigm sensitive to noise masking at favorable signal-to-noise ratios (SNRs). Two competing sentences differing in voice and context cues were presented against noise at SNRs of +2 and +6 dB. Listeners were asked to repeat back words from both competing sentences while prioritizing one of them. Recognition of the high-priority sentences was high and did not depend on the SNR. In contrast, recognition of the low-priority sentences was low and showed a significant SNR effect that was related to the listener's working memory capacity. This suggests that even subtle noise masking causes cognitive load in competing-talker situations.
Collapse
Affiliation(s)
- Hartmut Meister
- Jean-Uhrmacher-Institute for Clinical ENT-Research, University of Cologne, Geibelstrasse 29-31, D-50931 Cologne, Germany ,
| | - Sebastian Rählmann
- Jean-Uhrmacher-Institute for Clinical ENT-Research, University of Cologne, Geibelstrasse 29-31, D-50931 Cologne, Germany ,
| | - Martin Walger
- Clinic of Otorhinolaryngology, Head and Neck Surgery, University of Cologne, Kerpenerstr. 62, 50924 Cologne, Germany
| |
Collapse
|
14
|
Koelewijn T, Versfeld NJ, Kramer SE. Effects of attention on the speech reception threshold and pupil response of people with impaired and normal hearing. Hear Res 2017; 354:56-63. [PMID: 28869841 DOI: 10.1016/j.heares.2017.08.006] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/23/2017] [Revised: 08/21/2017] [Accepted: 08/25/2017] [Indexed: 11/26/2022]
Abstract
For people with hearing difficulties, following a conversation in a noisy environment requires substantial cognitive processing, which is often perceived as effortful. Recent studies with normal hearing (NH) listeners showed that the pupil dilation response, a measure of cognitive processing load, is affected by 'attention related' processes. How these processes affect the pupil dilation response for hearing impaired (HI) listeners remains unknown. Therefore, the current study investigated the effect of auditory attention on various pupil response parameters for 15 NH adults (median age 51 yrs.) and 15 adults with mild to moderate sensorineural hearing loss (median age 52 yrs.). Both groups listened to two different sentences presented simultaneously, one to each ear and partially masked by stationary noise. Participants had to repeat either both sentences or only one, for which they had to divide or focus attention, respectively. When repeating one sentence, the target sentence location (left or right) was either randomized or blocked across trials, which in the latter case allowed for a better spatial focus of attention. The speech-to-noise ratio was adjusted to yield about 50% sentences correct for each task and condition. NH participants had lower ('better') speech reception thresholds (SRT) than HI participants. The pupil measures showed no between-group effects, with the exception of a shorter peak latency for HI participants, which indicated a shorter processing time. Both groups showed higher SRTs and a larger pupil dilation response when two sentences were processed instead of one. Additionally, SRTs were higher and dilation responses were larger for both groups when the target location was randomized instead of fixed. We conclude that although HI participants could cope with less noise than the NH group, their ability to focus attention on a single talker, thereby improving SRTs and lowering cognitive processing load, was preserved. Shorter peak latencies could indicate that HI listeners adapt their listening strategy by not processing some information, which reduces processing time and thereby listening effort.
Collapse
Affiliation(s)
- Thomas Koelewijn
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and Amsterdam Public Health Research Institute, VU University Medical Center, Amsterdam, The Netherlands.
| | - Niek J Versfeld
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and Amsterdam Public Health Research Institute, VU University Medical Center, Amsterdam, The Netherlands
| | - Sophia E Kramer
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and Amsterdam Public Health Research Institute, VU University Medical Center, Amsterdam, The Netherlands
| |
Collapse
|
15
|
McCloy DR, Lau BK, Larson E, Pratt KAI, Lee AKC. Pupillometry shows the effort of auditory attention switching. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 141:2440. [PMID: 28464660 PMCID: PMC5848839 DOI: 10.1121/1.4979340] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Successful speech communication often requires selective attention to a target stream amidst competing sounds, as well as the ability to switch attention among multiple interlocutors. However, auditory attention switching negatively affects both target detection accuracy and reaction time, suggesting that attention switches carry a cognitive cost. Pupillometry is one method of assessing mental effort or cognitive load. Two experiments were conducted to determine whether the effort associated with attention switches is detectable in the pupillary response. In both experiments, pupil dilation, target detection sensitivity, and reaction time were measured; the task required listeners to either maintain or switch attention between two concurrent speech streams. Secondary manipulations explored whether switch-related effort would increase when auditory streaming was harder. In experiment 1, spatially distinct stimuli were degraded by simulating reverberation (compromising across-time streaming cues), and target-masker talker gender match was also varied. In experiment 2, diotic streams separable by talker voice quality and pitch were degraded by noise vocoding, and the time alloted for mid-trial attention switching was varied. All trial manipulations had some effect on target detection sensitivity and/or reaction time; however, only the attention-switching manipulation affected the pupillary response: greater dilation was observed in trials requiring switching attention between talkers.
Collapse
Affiliation(s)
- Daniel R McCloy
- Institute for Learning and Brain Sciences, University of Washington, 1715 Northeast Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Bonnie K Lau
- Institute for Learning and Brain Sciences, University of Washington, 1715 Northeast Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Eric Larson
- Institute for Learning and Brain Sciences, University of Washington, 1715 Northeast Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Katherine A I Pratt
- Institute for Learning and Brain Sciences, University of Washington, 1715 Northeast Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Adrian K C Lee
- Institute for Learning and Brain Sciences, University of Washington, 1715 Northeast Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| |
Collapse
|
16
|
Switching of auditory attention in "cocktail-party" listening: ERP evidence of cueing effects in younger and older adults. Brain Cogn 2016; 111:1-12. [PMID: 27814564 DOI: 10.1016/j.bandc.2016.09.006] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2016] [Revised: 06/28/2016] [Accepted: 09/13/2016] [Indexed: 11/23/2022]
Abstract
Verbal communication in a "cocktail-party situation" is a major challenge for the auditory system. In particular, changes in target speaker usually result in declined speech perception. Here, we investigated whether speech cues indicating a subsequent change in target speaker reduce the costs of switching in younger and older adults. We employed event-related potential (ERP) measures and a speech perception task, in which sequences of short words were simultaneously presented by four speakers. Changes in target speaker were either unpredictable or semantically cued by a word within the target stream. Cued changes resulted in a less decreased performance than uncued changes in both age groups. The ERP analysis revealed shorter latencies in the change-related N400 and late positive complex (LPC) after cued changes, suggesting an acceleration in context updating and attention switching. Thus, both younger and older listeners used semantic cues to prepare changes in speaker setting.
Collapse
|
17
|
|
18
|
Speech-in-Noise Tests and Supra-threshold Auditory Evoked Potentials as Metrics for Noise Damage and Clinical Trial Outcome Measures. Otol Neurotol 2016; 37:e295-302. [DOI: 10.1097/mao.0000000000001069] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
19
|
Koelewijn T, de Kluiver H, Shinn-Cunningham BG, Zekveld AA, Kramer SE. The pupil response reveals increased listening effort when it is difficult to focus attention. Hear Res 2015; 323:81-90. [PMID: 25732724 PMCID: PMC4632994 DOI: 10.1016/j.heares.2015.02.004] [Citation(s) in RCA: 70] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/22/2014] [Revised: 02/05/2015] [Accepted: 02/16/2015] [Indexed: 12/04/2022]
Abstract
Recent studies have shown that prior knowledge about where, when, and who is going to talk improves speech intelligibility. How related attentional processes affect cognitive processing load has not been investigated yet. In the current study, three experiments investigated how the pupil dilation response is affected by prior knowledge of target speech location, target speech onset, and who is going to talk. A total of 56 young adults with normal hearing participated. They had to reproduce a target sentence presented to one ear while ignoring a distracting sentence simultaneously presented to the other ear. The two sentences were independently masked by fluctuating noise. Target location (left or right ear), speech onset, and talker variability were manipulated in separate experiments by keeping these features either fixed during an entire block or randomized over trials. Pupil responses were recorded during listening and performance was scored after recall. The results showed an improvement in performance when the location of the target speech was fixed instead of randomized. Additionally, location uncertainty increased the pupil dilation response, which suggests that prior knowledge of location reduces cognitive load. Interestingly, the observed pupil responses for each condition were consistent with subjective reports of listening effort. We conclude that communicating in a dynamic environment like a cocktail party (where participants in competing conversations move unpredictably) requires substantial listening effort because of the demands placed on attentional processes.
Collapse
Affiliation(s)
- Thomas Koelewijn
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands.
| | - Hilde de Kluiver
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| | - Barbara G Shinn-Cunningham
- Department of Biomedical Engineering, Center for Computational Neuroscience and Neural Technology, Boston University, Boston, USA
| | - Adriana A Zekveld
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands; Linnaeus Centre HEAD, Department of Behavioral Sciences and Learning, Linköping University, Linköping, Sweden
| | - Sophia E Kramer
- Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery and EMGO Institute for Health and Care Research, VU University Medical Center, Amsterdam, The Netherlands
| |
Collapse
|
20
|
Getzmann S, Wascher E, Falkenstein M. What does successful speech-in-noise perception in aging depend on? Electrophysiological correlates of high and low performance in older adults. Neuropsychologia 2015; 70:43-57. [PMID: 25681737 DOI: 10.1016/j.neuropsychologia.2015.02.009] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2014] [Revised: 02/04/2015] [Accepted: 02/06/2015] [Indexed: 11/30/2022]
Abstract
Aging usually decreases the ability to understand language under difficult listening conditions. However, aging is also associated with increased between-subject variability. Here, we studied potential sources of inter-individual differences and investigated spoken language understanding of younger and older adults (age ranges 21-35 and 57-74 years, respectively) in a simulated "cocktail-party" scenario. A naturalistic "stock-price monitoring" task was employed in which prices of listed companies were simultaneously recited by four speakers at different locations in space. The participants responded when prices of a target company exceeded specific values, while ignoring all other companies. According to their individual performance levels three subgroups of participants were composed, consisting of 12 high-performing and 12 low-performing older adults, and 12 young adults matching the high-performing older group. The analysis of the event-related brain potentials indicated that all older adults showed delayed attentional control (indicated by a later P2) and reduced speech processing (indicated by a reduced N400), relative to the younger adults. High-performing older adults differed in increased allocation of attention and inhibitory control (indicated by a stronger P2-N2 complex) from their low-performing counterparts. The results are consistent with the idea of an adjustment of mental resources that could help compensating potential deficiencies in peripheral and central auditory processing.
Collapse
Affiliation(s)
- Stephan Getzmann
- Leibniz Research Centre for Working Environment and Human Factors, Ardeystraße 67, D-44139 Dortmund, Germany.
| | - Edmund Wascher
- Leibniz Research Centre for Working Environment and Human Factors, Ardeystraße 67, D-44139 Dortmund, Germany
| | - Michael Falkenstein
- Leibniz Research Centre for Working Environment and Human Factors, Ardeystraße 67, D-44139 Dortmund, Germany
| |
Collapse
|
21
|
At the interface of sensory and motor dysfunctions and Alzheimer's disease. Alzheimers Dement 2015; 11:70-98. [PMID: 25022540 PMCID: PMC4287457 DOI: 10.1016/j.jalz.2014.04.514] [Citation(s) in RCA: 379] [Impact Index Per Article: 42.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2013] [Revised: 03/13/2014] [Accepted: 04/09/2014] [Indexed: 11/21/2022]
Abstract
Recent evidence indicates that sensory and motor changes may precede the cognitive symptoms of Alzheimer's disease (AD) by several years and may signify increased risk of developing AD. Traditionally, sensory and motor dysfunctions in aging and AD have been studied separately. To ascertain the evidence supporting the relationship between age-related changes in sensory and motor systems and the development of AD and to facilitate communication between several disciplines, the National Institute on Aging held an exploratory workshop titled "Sensory and Motor Dysfunctions in Aging and AD." The scientific sessions of the workshop focused on age-related and neuropathologic changes in the olfactory, visual, auditory, and motor systems, followed by extensive discussion and hypothesis generation related to the possible links among sensory, cognitive, and motor domains in aging and AD. Based on the data presented and discussed at this workshop, it is clear that sensory and motor regions of the central nervous system are affected by AD pathology and that interventions targeting amelioration of sensory-motor deficits in AD may enhance patient function as AD progresses.
Collapse
|
22
|
MacPherson A, Akeroyd MA. Variations in the slope of the psychometric functions for speech intelligibility: a systematic survey. Trends Hear 2014; 18:18/0/2331216514537722. [PMID: 24906905 PMCID: PMC4227668 DOI: 10.1177/2331216514537722] [Citation(s) in RCA: 45] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
Although many studies have looked at the effects of different listening conditions on the intelligibility of speech, their analyses have often concentrated on changes to a single value on the psychometric function, namely, the threshold. Far less commonly has the slope of the psychometric function, that is, the rate at which intelligibility changes with level, been considered. The slope of the function is crucial because it is the slope, rather than the threshold, that determines the improvement in intelligibility caused by any given improvement in signal-to-noise ratio by, for instance, a hearing aid. The aim of the current study was to systematically survey and reanalyze the psychometric function data available in the literature in an attempt to quantify the range of slope changes across studies and to identify listening conditions that affect the slope of the psychometric function. The data for 885 individual psychometric functions, taken from 139 different studies, were fitted with a common logistic equation from which the slope was calculated. Large variations in slope across studies were found, with slope values ranging from as shallow as 1% per dB to as steep as 44% per dB (median = 6.6% per dB), suggesting that the perceptual benefit offered by an improvement in signal-to-noise ratio depends greatly on listening environment. The type and number of maskers used were found to be major factors on the value of the slope of the psychometric function while other minor effects of target predictability, target corpus, and target/masker similarity were also found.
Collapse
Affiliation(s)
- Alexandra MacPherson
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, Glasgow, UK School of Psychological Sciences & Health, University of Strathclyde, Glasgow, UK
| | - Michael A Akeroyd
- MRC/CSO Institute of Hearing Research-Scottish Section, Glasgow Royal Infirmary, Glasgow, UK
| |
Collapse
|
23
|
The pupil response is sensitive to divided attention during speech processing. Hear Res 2014; 312:114-20. [PMID: 24709275 PMCID: PMC4634867 DOI: 10.1016/j.heares.2014.03.010] [Citation(s) in RCA: 58] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/26/2013] [Revised: 03/12/2014] [Accepted: 03/24/2014] [Indexed: 11/23/2022]
Abstract
Dividing attention over two streams of speech strongly decreases performance compared to focusing on only one. How divided attention affects cognitive processing load as indexed with pupillometry during speech recognition has so far not been investigated. In 12 young adults the pupil response was recorded while they focused on either one or both of two sentences that were presented dichotically and masked by fluctuating noise across a range of signal-to-noise ratios. In line with previous studies, the performance decreases when processing two target sentences instead of one. Additionally, dividing attention to process two sentences caused larger pupil dilation and later peak pupil latency than processing only one. This suggests an effect of attention on cognitive processing load (pupil dilation) during speech processing in noise.
Collapse
|
24
|
Koelewijn T, Zekveld AA, Festen JM, Kramer SE. The influence of informational masking on speech perception and pupil response in adults with hearing impairment. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2014; 135:1596-606. [PMID: 24606294 DOI: 10.1121/1.4863198] [Citation(s) in RCA: 43] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/21/2023]
Abstract
A recent pupillometry study on adults with normal hearing indicates that the pupil response during speech perception (cognitive processing load) is strongly affected by the type of speech masker. The current study extends these results by recording the pupil response in 32 participants with hearing impairment (mean age 59 yr) while they were listening to sentences masked by fluctuating noise or a single-talker. Efforts were made to improve audibility of all sounds by means of spectral shaping. Additionally, participants performed tests measuring verbal working memory capacity, inhibition of interfering information in working memory, and linguistic closure. The results showed worse speech reception thresholds for speech masked by single-talker speech compared to fluctuating noise. In line with previous results for participants with normal hearing, the pupil response was larger when listening to speech masked by a single-talker compared to fluctuating noise. Regression analysis revealed that larger working memory capacity and better inhibition of interfering information related to better speech reception thresholds, but these variables did not account for inter-individual differences in the pupil response. In conclusion, people with hearing impairment show more cognitive load during speech processing when there is interfering speech compared to fluctuating noise.
Collapse
Affiliation(s)
- Thomas Koelewijn
- Department of Otolaryngology-Head and Neck Surgery, section Audiology and EMGO Institute for Health and Care Research, VU University Medical Center, Boelelaan 1118, 1081 HZ Amsterdam, The Netherlands
| | - Adriana A Zekveld
- Department of Otolaryngology-Head and Neck Surgery, section Audiology and EMGO Institute for Health and Care Research, VU University Medical Center, Boelelaan 1118, 1081 HZ Amsterdam, The Netherlands
| | - Joost M Festen
- Department of Otolaryngology-Head and Neck Surgery, section Audiology and EMGO Institute for Health and Care Research, VU University Medical Center, Boelelaan 1118, 1081 HZ Amsterdam, The Netherlands
| | - Sophia E Kramer
- Department of Otolaryngology-Head and Neck Surgery, section Audiology and EMGO Institute for Health and Care Research, VU University Medical Center, Boelelaan 1118, 1081 HZ Amsterdam, The Netherlands
| |
Collapse
|
25
|
Gygi B, Shafiro V. Spatial and temporal modifications of multitalker speech can improve speech perception in older adults. Hear Res 2014; 310:76-86. [PMID: 24530609 DOI: 10.1016/j.heares.2014.01.009] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/17/2013] [Revised: 01/23/2014] [Accepted: 01/28/2014] [Indexed: 11/16/2022]
Abstract
Speech perception in multitalker environments often requires listeners to divide attention among several concurrent talkers before focusing on one talker with pertinent information. Such attentionally demanding tasks are particularly difficult for older adults due both to age-related hearing loss (presbacusis) and general declines in attentional processing and associated cognitive abilities. This study investigated two signal-processing techniques that have been suggested as a means of improving speech perception accuracy of older adults: time stretching and spatial separation of target talkers. Stimuli in each experiment comprised 2-4 fixed-form utterances in which listeners were asked to consecutively 1) detect concurrently spoken keywords in the beginning of the utterance (divided attention); and, 2) identify additional keywords from only one talker at the end of the utterance (selective attention). In Experiment 1, the overall tempo of each utterance was unaltered or slowed down by 25%; in Experiment 2 the concurrent utterances were spatially coincident or separated across a 180-degree hemifield. Both manipulations improved performance for elderly adults with age-appropriate hearing on both tasks. Increasing the divided attention load by attending to more concurrent keywords had a marked negative effect on performance of the selective attention task only when the target talker was identified by a keyword, but not by spatial location. These findings suggest that the temporal and spatial modifications of multitalker speech improved perception of multitalker speech primarily by reducing competition among cognitive resources required to perform attentionally demanding tasks.
Collapse
Affiliation(s)
- Brian Gygi
- National Institute for Health Research, Nottingham Hearing Biomedical Research Unit, Nottingham, UK; Otology and Hearing Group, Division of Clinical Neuroscience, School of Medicine, University of Nottingham, Nottingham NG7 2UH, UK; Veterans Affairs Northern California Health Care System, Martinez, CA, USA.
| | | |
Collapse
|
26
|
Gallun FJ, Diedesch AC, Kampel SD, Jakien KM. Independent impacts of age and hearing loss on spatial release in a complex auditory environment. Front Neurosci 2013; 7:252. [PMID: 24391535 PMCID: PMC3870327 DOI: 10.3389/fnins.2013.00252] [Citation(s) in RCA: 85] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2013] [Accepted: 12/06/2013] [Indexed: 11/26/2022] Open
Abstract
Listeners in complex auditory environments can benefit from the ability to use a variety of spatial and spectrotemporal cues for sound source segregation. Probing these abilities is an essential part of gaining a more complete understanding of why listeners differ in navigating the auditory environment. Two fundamental processes that can impact the auditory systems of individual listeners are aging and hearing loss. One difficulty with uncovering the independent effects of age and hearing loss on spatial release is the commonly observed phenomenon of age-related hearing loss. In order to reveal the effects of aging on spatial hearing, it is essential to develop testing methods that reduce the influence of hearing loss on the outcomes. The statistical power needed for such testing generally requires a larger number of participants than can easily be tested using traditional behavioral methods. This work describes the development and validation of a rapid method by which listeners can be categorized in terms of their ability to use spatial and spectrotemporal cues to separate competing speech streams. Results show that when age and audibility are not covarying, age alone can be shown to substantially reduce spatial release from masking. These data support the hypothesis that aging, independent of an individual's hearing threshold, can result in changes in the cortical and/or subcortical structures essential for spatial hearing.
Collapse
Affiliation(s)
- Frederick J Gallun
- Department of Veterans Affairs, National Center for Rehabilitative Auditory Research, Portland VA Medical Center Portland, OR, USA ; Otolaryngology/Head and Neck Surgery, Oregon Health and Science University Portland, OR, USA
| | - Anna C Diedesch
- Hearing and Speech Sciences, Vanderbilt University Nashville, TN, USA
| | - Sean D Kampel
- Department of Veterans Affairs, National Center for Rehabilitative Auditory Research, Portland VA Medical Center Portland, OR, USA
| | - Kasey M Jakien
- Otolaryngology/Head and Neck Surgery, Oregon Health and Science University Portland, OR, USA
| |
Collapse
|
27
|
Woods WS, Kalluri S, Pentony S, Nooraei N. Predicting the effect of hearing loss and audibility on amplified speech reception in a multi-talker listening scenario. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:4268-78. [PMID: 23742377 DOI: 10.1121/1.4803859] [Citation(s) in RCA: 22] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Auditive and cognitive influences on speech perception in a complex situation were investigated in listeners with normal hearing (NH) and hearing loss (HL). The speech corpus used was the Nonsense-Syllable Response Measure [NSRM; Woods and Kalluri, (2010). International Hearing Aid Research Conference, pp. 40-41], a 12-talker corpus which combines 154 nonsense syllables with 8 different carrier phrases. Listeners heard NSRM sentences in quiet, background noise, and in background noise plus other "jammer" NSRM sentences. All stimuli were linearly amplified. A "proficiency" value, determined from the results in quiet and the quiet-condition speech intelligibility index (SII), was used with the SII in predicting results in the other conditions. Results for nine of ten NH subjects were well-predicted (within the limits of binomial variability) in the noise condition, as were eight of these subjects in the noise-plus-jammers condition. All 16 HL results were well-predicted in the noise condition, as were 9 of the HL in the noise-plus-jammers condition. Hierarchical regression partialling out the effects of age found proficiency in noise-plus-jammers significantly correlated with results of "trail-making" tests, thought to index processing speed and attention-deployment ability, and proficiency in quiet and noise was found significantly correlated with results from a backward digit-span memory test.
Collapse
Affiliation(s)
- William S Woods
- Starkey Hearing Research Center, 2150 Shattuck Avenue, Suite 408, Berkeley, California 94704, USA.
| | | | | | | |
Collapse
|
28
|
Schwartz AH, Shinn-Cunningham BG. Effects of dynamic range compression on spatial selective auditory attention in normal-hearing listeners. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2013; 133:2329-2339. [PMID: 23556599 PMCID: PMC3631248 DOI: 10.1121/1.4794386] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/05/2012] [Revised: 02/13/2013] [Accepted: 02/14/2013] [Indexed: 05/31/2023]
Abstract
Many hearing aids introduce compressive gain to accommodate the reduced dynamic range that often accompanies hearing loss. However, natural sounds produce complicated temporal dynamics in hearing aid compression, as gain is driven by whichever source dominates at a given moment. Moreover, independent compression at the two ears can introduce fluctuations in interaural level differences (ILDs) important for spatial perception. While independent compression can interfere with spatial perception of sound, it does not always interfere with localization accuracy or speech identification. Here, normal-hearing listeners reported a target message played simultaneously with two spatially separated masker messages. We measured the amount of spatial separation required between the target and maskers for subjects to perform at threshold in this task. Fast, syllabic compression that was independent at the two ears increased the required spatial separation, but linking the compressors to provide identical gain to both ears (preserving ILDs) restored much of the deficit caused by fast, independent compression. Effects were less clear for slower compression. Percent-correct performance was lower with independent compression, but only for small spatial separations. These results may help explain differences in previous reports of the effect of compression on spatial perception of sound.
Collapse
Affiliation(s)
- Andrew H Schwartz
- Harvard/Massachusetts Institute of Technology, Speech and Hearing Bioscience and Technology Program, Cambridge, Massachusetts 02139, USA
| | | |
Collapse
|
29
|
Kong L, Michalka SW, Rosen ML, Sheremata SL, Swisher JD, Shinn-Cunningham BG, Somers DC. Auditory spatial attention representations in the human cerebral cortex. Cereb Cortex 2012. [PMID: 23180753 DOI: 10.1093/cercor/bhs359] [Citation(s) in RCA: 54] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022] Open
Abstract
Auditory spatial attention serves important functions in auditory source separation and selection. Although auditory spatial attention mechanisms have been generally investigated, the neural substrates encoding spatial information acted on by attention have not been identified in the human neocortex. We performed functional magnetic resonance imaging experiments to identify cortical regions that support auditory spatial attention and to test 2 hypotheses regarding the coding of auditory spatial attention: 1) auditory spatial attention might recruit the visuospatial maps of the intraparietal sulcus (IPS) to create multimodal spatial attention maps; 2) auditory spatial information might be encoded without explicit cortical maps. We mapped visuotopic IPS regions in individual subjects and measured auditory spatial attention effects within these regions of interest. Contrary to the multimodal map hypothesis, we observed that auditory spatial attentional modulations spared the visuotopic maps of IPS; the parietal regions activated by auditory attention lacked map structure. However, multivoxel pattern analysis revealed that the superior temporal gyrus and the supramarginal gyrus contained significant information about the direction of spatial attention. These findings support the hypothesis that auditory spatial information is coded without a cortical map representation. Our findings suggest that audiospatial and visuospatial attention utilize distinctly different spatial coding schemes.
Collapse
|
30
|
Spatial selective auditory attention in the presence of reverberant energy: individual differences in normal-hearing listeners. J Assoc Res Otolaryngol 2010; 12:395-405. [PMID: 21128091 DOI: 10.1007/s10162-010-0254-z] [Citation(s) in RCA: 50] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2010] [Accepted: 11/16/2010] [Indexed: 10/18/2022] Open
Abstract
Listeners can selectively attend to a desired target by directing attention to known target source features, such as location or pitch. Reverberation, however, reduces the reliability of the cues that allow a target source to be segregated and selected from a sound mixture. Given this, it is likely that reverberant energy interferes with selective auditory attention. Anecdotal reports suggest that the ability to focus spatial auditory attention degrades even with early aging, yet there is little evidence that middle-aged listeners have behavioral deficits on tasks requiring selective auditory attention. The current study was designed to look for individual differences in selective attention ability and to see if any such differences correlate with age. Normal-hearing adults, ranging in age from 18 to 55 years, were asked to report a stream of digits located directly ahead in a simulated rectangular room. Simultaneous, competing masker digit streams were simulated at locations 15° left and right of center. The level of reverberation was varied to alter task difficulty by interfering with localization cues (increasing localization blur). Overall, performance was best in the anechoic condition and worst in the high-reverberation condition. Listeners nearly always reported a digit from one of the three competing streams, showing that reverberation did not render the digits unintelligible. Importantly, inter-subject differences were extremely large. These differences, however, were not significantly correlated with age, memory span, or hearing status. These results show that listeners with audiometrically normal pure tone thresholds differ in their ability to selectively attend to a desired source, a task important in everyday communication. Further work is necessary to determine if these differences arise from differences in peripheral auditory function or in more central function.
Collapse
|
31
|
Kidd G, Mason CR, Best V, Marrone N. Stimulus factors influencing spatial release from speech-on-speech masking. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2010; 128:1965-78. [PMID: 20968368 PMCID: PMC2981113 DOI: 10.1121/1.3478781] [Citation(s) in RCA: 67] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
This study examined spatial release from masking (SRM) when a target talker was masked by competing talkers or by other types of sounds. The focus was on the role of interaural time differences (ITDs) and time-varying interaural level differences (ILDs) under conditions varying in the strength of informational masking (IM). In the first experiment, a target talker was masked by two other talkers that were either colocated with the target or were symmetrically spatially separated from the target with the stimuli presented through loudspeakers. The sounds were filtered into different frequency regions to restrict the available interaural cues. The largest SRM occurred for the broadband condition followed by a low-pass condition. However, even the highest frequency bandpass-filtered condition (3-6 kHz) yielded a significant SRM. In the second experiment the stimuli were presented via earphones. The listeners identified the speech of a target talker masked by one or two other talkers or noises when the maskers were colocated with the target or were perceptually separated by ITDs. The results revealed a complex pattern of masking in which the factors affecting performance in colocated and spatially separated conditions are to a large degree independent.
Collapse
Affiliation(s)
- Gerald Kidd
- Department of Speech, Language and Hearing Sciences, and Hearing Research Center, Boston University, Boston, Massachusetts 02215, USA
| | | | | | | |
Collapse
|