1
|
Xu S, Zhang H, Fan J, Jiang X, Zhang M, Guan J, Ding H, Zhang Y. Auditory Challenges and Listening Effort in School-Age Children With Autism: Insights From Pupillary Dynamics During Speech-in-Noise Perception. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024; 67:2410-2453. [PMID: 38861391 DOI: 10.1044/2024_jslhr-23-00553] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2024]
Abstract
PURPOSE This study aimed to investigate challenges in speech-in-noise (SiN) processing faced by school-age children with autism spectrum conditions (ASCs) and their impact on listening effort. METHOD Participants, including 23 Mandarin-speaking children with ASCs and 19 age-matched neurotypical (NT) peers, underwent sentence recognition tests in both quiet and noisy conditions, with a speech-shaped steady-state noise masker presented at 0-dB signal-to-noise ratio in the noisy condition. Recognition accuracy rates and task-evoked pupil responses were compared to assess behavioral performance and listening effort during auditory tasks. RESULTS No main effect of group was found on accuracy rates. Instead, significant effects emerged for autistic trait scores, listening conditions, and their interaction, indicating that higher trait scores were associated with poorer performance in noise. Pupillometric data revealed significantly larger and earlier peak dilations, along with more varied pupillary dynamics in the ASC group relative to the NT group, especially under noisy conditions. Importantly, the ASC group's peak dilation in quiet mirrored that of the NT group in noise. However, the ASC group consistently exhibited reduced mean dilations than the NT group. CONCLUSIONS Pupillary responses suggest a different resource allocation pattern in ASCs: An initial sharper and larger dilation may signal an intense, narrowed resource allocation, likely linked to heightened arousal, engagement, and cognitive load, whereas a subsequent faster tail-off may indicate a greater decrease in resource availability and engagement, or a quicker release of arousal and cognitive load. The presence of noise further accentuates this pattern. This highlights the unique SiN processing challenges children with ASCs may face, underscoring the importance of a nuanced, individual-centric approach for interventions and support.
Collapse
Affiliation(s)
- Suyun Xu
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- National Research Centre for Language and Well-Being, Shanghai, China
| | - Hua Zhang
- Department of Child and Adolescent Psychiatry, Shanghai Mental Health Center, Shanghai Jiao Tong University School of Medicine, China
| | - Juan Fan
- Department of Child and Adolescent Psychiatry, Shanghai Mental Health Center, Shanghai Jiao Tong University School of Medicine, China
| | - Xiaoming Jiang
- Institute of Linguistics, Shanghai International Studies University, China
| | - Minyue Zhang
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- National Research Centre for Language and Well-Being, Shanghai, China
| | | | - Hongwei Ding
- Speech-Language-Hearing Center, School of Foreign Languages, Shanghai Jiao Tong University, China
- National Research Centre for Language and Well-Being, Shanghai, China
| | - Yang Zhang
- Department of Speech-Language-Hearing Sciences and Masonic Institute for the Developing Brain, University of Minnesota, Minneapolis
| |
Collapse
|
2
|
Kuk F, Slugocki C, Korhonen P. Characteristics of the quick repeat-recall test (Q-RRT). Int J Audiol 2024; 63:482-490. [PMID: 37615510 DOI: 10.1080/14992027.2023.2245969] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Accepted: 08/03/2023] [Indexed: 08/25/2023]
Abstract
OBJECTIVE To determine the reliability and validity of the Quick Repeat-Recall Test (Q-RRT). DESIGN Within-subject repeated measures. Participants completed the Q-RRT at a speech level of 75 dB SPL in 2 noise configurations at signal-to-noise ratios of 5, 10, and 15 dB in a counterbalanced order, along with the full-RRT. The Q-RRT was repeated after 1-3 weeks to estimate within-session and between-session reliability. Participants also completed external validation measures relating speech-in-noise ability (Hearing-In-Noise Test, HINT), memory (Reading Span Test, RST; Forward Digit Span Test, DST-F), and noise acceptance (Tracking of Noise Tolerance, TNT). STUDY SAMPLE Twenty-five normal-hearing listeners (mean age = 63.9 years). RESULTS Group Repeat and Recall scores were similar between the full-RRT and the Q-RRT. However, Listening Effort and Tolerable Time ratings were significantly poorer for the full-RRT than the Q-RRT. The average repeat scores correlated with the HINT scores, while the recall scores correlated with the DST-F and RST scores. Tolerable Time ratings also correlated with the TNT scores. Within-session and between-session test-retest reliability using intraclass correlation coefficients were fair-to-excellent (0.41-0.91) depending on the outcome measures. CONCLUSIONS The Q-RRT is a reliable and valid measure of speech-in-noise ability, working memory, listening effort, and noise acceptance.
Collapse
Affiliation(s)
- Francis Kuk
- WSA Office of Research in Clinical Amplification (ORCA-USA), Lisle, IL, USA
| | | | - Petri Korhonen
- WSA Office of Research in Clinical Amplification (ORCA-USA), Lisle, IL, USA
| |
Collapse
|
3
|
Mechtenberg H, Giorio C, Myers EB. Pupil Dilation Reflects Perceptual Priorities During a Receptive Speech Task. Ear Hear 2024; 45:425-440. [PMID: 37882091 PMCID: PMC10868674 DOI: 10.1097/aud.0000000000001438] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2022] [Accepted: 09/01/2023] [Indexed: 10/27/2023]
Abstract
OBJECTIVES The listening demand incurred by speech perception fluctuates in normal conversation. At the acoustic-phonetic level, natural variation in pronunciation acts as speedbumps to accurate lexical selection. Any given utterance may be more or less phonetically ambiguous-a problem that must be resolved by the listener to choose the correct word. This becomes especially apparent when considering two common speech registers-clear and casual-that have characteristically different levels of phonetic ambiguity. Clear speech prioritizes intelligibility through hyperarticulation which results in less ambiguity at the phonetic level, while casual speech tends to have a more collapsed acoustic space. We hypothesized that listeners would invest greater cognitive resources while listening to casual speech to resolve the increased amount of phonetic ambiguity, as compared with clear speech. To this end, we used pupillometry as an online measure of listening effort during perception of clear and casual continuous speech in two background conditions: quiet and noise. DESIGN Forty-eight participants performed a probe detection task while listening to spoken, nonsensical sentences (masked and unmasked) while recording pupil size. Pupil size was modeled using growth curve analysis to capture the dynamics of the pupil response as the sentence unfolded. RESULTS Pupil size during listening was sensitive to the presence of noise and speech register (clear/casual). Unsurprisingly, listeners had overall larger pupil dilations during speech perception in noise, replicating earlier work. The pupil dilation pattern for clear and casual sentences was considerably more complex. Pupil dilation during clear speech trials was slightly larger than for casual speech, across quiet and noisy backgrounds. CONCLUSIONS We suggest that listener motivation could explain the larger pupil dilations to clearly spoken speech. We propose that, bounded by the context of this task, listeners devoted more resources to perceiving the speech signal with the greatest acoustic/phonetic fidelity. Further, we unexpectedly found systematic differences in pupil dilation preceding the onset of the spoken sentences. Together, these data demonstrate that the pupillary system is not merely reactive but also adaptive-sensitive to both task structure and listener motivation to maximize accurate perception in a limited resource system.
Collapse
Affiliation(s)
- Hannah Mechtenberg
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, USA
| | - Cristal Giorio
- Department of Psychology, Pennsylvania State University, State College, Pennsylvania, USA
| | - Emily B. Myers
- Department of Psychological Sciences, University of Connecticut, Storrs, Connecticut, USA
- Department of Speech, Language and Hearing Sciences, University of Connecticut, Storrs, Connecticut, USA
| |
Collapse
|
4
|
Villard S, Perrachione TK, Lim SJ, Alam A, Kidd G. Energetic and informational masking place dissociable demands on listening effort: Evidence from simultaneous electroencephalography and pupillometrya). THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 154:1152-1167. [PMID: 37610284 PMCID: PMC10449482 DOI: 10.1121/10.0020539] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/21/2022] [Revised: 07/09/2023] [Accepted: 07/14/2023] [Indexed: 08/24/2023]
Abstract
The task of processing speech masked by concurrent speech/noise can pose a substantial challenge to listeners. However, performance on such tasks may not directly reflect the amount of listening effort they elicit. Changes in pupil size and neural oscillatory power in the alpha range (8-12 Hz) are prominent neurophysiological signals known to reflect listening effort; however, measurements obtained through these two approaches are rarely correlated, suggesting that they may respond differently depending on the specific cognitive demands (and, by extension, the specific type of effort) elicited by specific tasks. This study aimed to compare changes in pupil size and alpha power elicited by different types of auditory maskers (highly confusable intelligible speech maskers, speech-envelope-modulated speech-shaped noise, and unmodulated speech-shaped noise maskers) in young, normal-hearing listeners. Within each condition, the target-to-masker ratio was set at the participant's individually estimated 75% correct point on the psychometric function. The speech masking condition elicited a significantly greater increase in pupil size than either of the noise masking conditions, whereas the unmodulated noise masking condition elicited a significantly greater increase in alpha oscillatory power than the speech masking condition, suggesting that the effort needed to solve these respective tasks may have different neural origins.
Collapse
Affiliation(s)
- Sarah Villard
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Tyler K Perrachione
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Sung-Joo Lim
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Ayesha Alam
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| | - Gerald Kidd
- Department of Speech, Language, and Hearing Sciences, Boston University, Boston, Massachusetts 02215, USA
| |
Collapse
|
5
|
Kaufman M, Zion Golumbic E. Listening to two speakers: Capacity and tradeoffs in neural speech tracking during Selective and Distributed Attention. Neuroimage 2023; 270:119984. [PMID: 36854352 DOI: 10.1016/j.neuroimage.2023.119984] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Revised: 02/06/2023] [Accepted: 02/24/2023] [Indexed: 02/27/2023] Open
Abstract
Speech comprehension is severely compromised when several people talk at once, due to limited perceptual and cognitive resources. In such circumstances, top-down attention mechanisms can actively prioritize processing of task-relevant speech. However, behavioral and neural evidence suggest that this selection is not exclusive, and the system may have sufficient capacity to process additional speech input as well. Here we used a data-driven approach to contrast two opposing hypotheses regarding the system's capacity to co-represent competing speech: Can the brain represent two speakers equally or is the system fundamentally limited, resulting in tradeoffs between them? Neural activity was measured using magnetoencephalography (MEG) as human participants heard concurrent speech narratives and engaged in two tasks: Selective Attention, where only one speaker was task-relevant and Distributed Attention, where both speakers were equally relevant. Analysis of neural speech-tracking revealed that both tasks engaged a similar network of brain regions involved in auditory processing, attentional control and speech processing. Interestingly, during both Selective and Distributed Attention the neural representation of competing speech showed a bias towards one speaker. This is in line with proposed 'bottlenecks' for co-representation of concurrent speech and suggests that good performance on distributed attention tasks may be achieved by toggling attention between speakers over time.
Collapse
Affiliation(s)
- Maya Kaufman
- The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, Israel
| | - Elana Zion Golumbic
- The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, Israel.
| |
Collapse
|
6
|
Neagu MB, Kressner AA, Relaño-Iborra H, Bækgaard P, Dau T, Wendt D. Investigating the Reliability of Pupillometry as a Measure of
Individualized Listening Effort. Trends Hear 2023; 27:23312165231153288. [PMCID: PMC9947699 DOI: 10.1177/23312165231153288] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/24/2023] Open
Abstract
Recordings of the pupillary response have been used in numerous studies to assess
listening effort during a speech-in-noise task. Most studies focused on averaged
responses across listeners, whereas less is known about pupil dilation as an
indicator of the individuals’ listening effort. The present study investigated
the reliability of several pupil features as potential indicators of individual
listening effort and the impact of different normalization procedures on the
reliability. The pupil diameters of 31 normal-hearing listeners were recorded
during multiple visits while performing a speech-in-noise task. The
signal-to-noise ratios (SNRs) of the stimuli ranged from
−12 dB to
+4 dB. All listeners
were measured twice at separate visits, and 11 were re-tested at a third visit.
To examine the reliability of the pupil responses across visits, the intraclass
correlation coefficient was applied to the peak and mean pupil dilation and to
the temporal features of the pupil response, extracted using growth curve
analysis. The reliability of the pupillary response was assessed in relation to
SNR and different normalization procedures over multiple visits. The most
reliable pupil features were the traditional mean and peak pupil dilation. The
highest reliability results were obtained when the data were baseline-corrected
and normalized to the individual pupil response range across all visits.
Moreover, the present study results showed only a minor impact of the SNR and
the number of visits on the reliability of the pupil response. Overall, the
results may provide an important basis for developing a standardized test for
pupillometry in the clinic.
Collapse
Affiliation(s)
- Mihaela-Beatrice Neagu
- Department of Health Technology, DTU Hearing Systems, Denmark,Mihaela-Beatrice Neagu, Department of
Health Technology, DTU Hearing Systems, Denmark.
| | - Abigail A. Kressner
- Department of Health Technology, DTU Hearing Systems, Denmark,Copenhagen Hearing and Balance Centre, Rigshospitalet, Copenhagen
University Hospital, Denmark
| | - Helia Relaño-Iborra
- Department of Health Technology, DTU Hearing Systems, Denmark,Department of Applied Mathematics and Computer Science, DTU Cognitive systems, Denmark
| | - Per Bækgaard
- Department of Applied Mathematics and Computer Science, DTU Cognitive systems, Denmark
| | - Torsten Dau
- Department of Health Technology, DTU Hearing Systems, Denmark,Copenhagen Hearing and Balance Centre, Rigshospitalet, Copenhagen
University Hospital, Denmark
| | - Dorothea Wendt
- Department of Health Technology, DTU Hearing Systems, Denmark,Eriksholm Research Centre, Denmark
| |
Collapse
|
7
|
Wasiuk PA, Buss E, Oleson JJ, Calandruccio L. Predicting speech-in-speech recognition: Short-term audibility, talker sex, and listener factors. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 152:3010. [PMID: 36456289 DOI: 10.1121/10.0015228] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Accepted: 11/01/2022] [Indexed: 06/17/2023]
Abstract
Speech-in-speech recognition can be challenging, and listeners vary considerably in their ability to accomplish this complex auditory-cognitive task. Variability in performance can be related to intrinsic listener factors as well as stimulus factors associated with energetic and informational masking. The current experiments characterized the effects of short-term audibility of the target, differences in target and masker talker sex, and intrinsic listener variables on sentence recognition in two-talker speech and speech-shaped noise. Participants were young adults with normal hearing. Each condition included the adaptive measurement of speech reception thresholds, followed by testing at a fixed signal-to-noise ratio (SNR). Short-term audibility for each keyword was quantified using a computational glimpsing model for target+masker mixtures. Scores on a psychophysical task of auditory stream segregation predicted speech recognition, with stronger effects for speech-in-speech than speech-in-noise. Both speech-in-speech and speech-in-noise recognition depended on the proportion of audible glimpses available in the target+masker mixture, even across stimuli presented at the same global SNR. Short-term audibility requirements varied systematically across stimuli, providing an estimate of the greater informational masking for speech-in-speech than speech-in-noise recognition and quantifying informational masking for matched and mismatched talker sex.
Collapse
Affiliation(s)
- Peter A Wasiuk
- Department of Psychological Sciences, 11635 Euclid Avenue, Case Western Reserve University, Cleveland, Ohio 44106, USA
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, 170 Manning Drive, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA
| | - Jacob J Oleson
- Department of Biostatistics, 145 North Riverside Drive, University of Iowa, Iowa City, Iowa 52242, USA
| | - Lauren Calandruccio
- Department of Psychological Sciences, 11635 Euclid Avenue, Case Western Reserve University, Cleveland, Ohio 44106, USA
| |
Collapse
|
8
|
Meng Y, Chen F, Feng Y, Peng G, Zheng W. Age-Related Differences of Mandarin Tone and Consonant Aspiration Perception in Babble Noise. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:3438-3451. [PMID: 36044891 DOI: 10.1044/2022_jslhr-21-00564] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
PURPOSE This study investigated the categorical perception of Mandarin tones and consonant aspiration contrasts in babble noise among adults and adolescents aged 12-14 years, and explored the association between working memory and categorical perception. METHOD Twenty-four adults and 20 adolescents with Mandarin as their native language were recruited. Their performances of phonemic identification and discrimination in babble noise and quiet conditions, digit span tasks, and nonword repetition were assessed. RESULTS Results indicated that, firstly, in the noise condition, both adults and adolescents showed wider boundary widths and lower between-category accuracies when perceiving aspiration of consonants than in the quiet condition, and the categorical boundary of tone perception in adolescents showed a transitional tendency toward Tone 1. Secondly, discrimination of consonant aspiration in adolescents needed to be further developed. Lastly, the accuracy of nonword repetition in adolescents was lower than that in adults, and adults with better auditory verbal working memory had better performance on tone perception. CONCLUSIONS Our results provided evidence that tone perception is acquired easier than consonant aspiration perception, and tone perception is more robust and less susceptible to noise interference. Categorical perception performance relates to the capacity and utilization of auditory verbal working memory in some ways.
Collapse
Affiliation(s)
- Yaru Meng
- Department of Chinese Language and Literature, East China Normal University, Shanghai
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, China
| | - Fei Chen
- School of Foreign Languages, Hunan University, Changsha, China
| | - Yan Feng
- School of Foreign Studies, Nanjing University of Science and Technology, Nanjing, China
- Research Centre for Language, Cognition, and Neuroscience, Department of Chinese and Bilingual Studies, The Hong Kong Polytechnic University, China
| | - Gang Peng
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, China
- Research Centre for Language, Cognition, and Neuroscience, Department of Chinese and Bilingual Studies, The Hong Kong Polytechnic University, China
| | - Wei Zheng
- Department of Chinese Language and Literature, East China Normal University, Shanghai
| |
Collapse
|
9
|
Strauch C, Wang CA, Einhäuser W, Van der Stigchel S, Naber M. Pupillometry as an integrated readout of distinct attentional networks. Trends Neurosci 2022; 45:635-647. [PMID: 35662511 DOI: 10.1016/j.tins.2022.05.003] [Citation(s) in RCA: 59] [Impact Index Per Article: 29.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Revised: 04/15/2022] [Accepted: 05/09/2022] [Indexed: 10/18/2022]
Abstract
The course of pupillary constriction and dilation provides an easy-to-access, inexpensive, and noninvasive readout of brain activity. We propose a new taxonomy of factors affecting the pupil and link these to associated neural underpinnings in an ascending hierarchy. In addition to two well-established low-level factors (light level and focal distance), we suggest two further intermediate-level factors, alerting and orienting, and a higher-level factor, executive functioning. Alerting, orienting, and executive functioning - including their respective underlying neural circuitries - overlap with the three principal attentional networks, making pupil size an integrated readout of distinct states of attention. As a now widespread technique, pupillometry is ready to provide meaningful applications and constitutes a viable part of the psychophysiological toolbox.
Collapse
Affiliation(s)
- Christoph Strauch
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands.
| | - Chin-An Wang
- Institute of Cognitive Neuroscience, National Central University, Taoyuan City, Taiwan; Cognitive Intelligence and Precision Healthcare Center, National Central University, Taoyuan City, Taiwan
| | - Wolfgang Einhäuser
- Physics of Cognition Group, Chemnitz University of Technology, Chemnitz, Germany
| | | | - Marnix Naber
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
10
|
Dingemanse G, Goedegebure A. Listening Effort in Cochlear Implant Users: The Effect of Speech Intelligibility, Noise Reduction Processing, and Working Memory Capacity on the Pupil Dilation Response. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2022; 65:392-404. [PMID: 34898265 DOI: 10.1044/2021_jslhr-21-00230] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
PURPOSE This study aimed to evaluate the effect of speech recognition performance, working memory capacity (WMC), and a noise reduction algorithm (NRA) on listening effort as measured with pupillometry in cochlear implant (CI) users while listening to speech in noise. METHOD Speech recognition and pupil responses (peak dilation, peak latency, and release of dilation) were measured during a speech recognition task at three speech-to-noise ratios (SNRs) with an NRA in both on and off conditions. WMC was measured with a reading span task. Twenty experienced CI users participated in this study. RESULTS With increasing SNR and speech recognition performance, (a) the peak pupil dilation decreased by only a small amount, (b) the peak latency decreased, and (c) the release of dilation after the sentences increased. The NRA had no effect on speech recognition in noise or on the peak or latency values of the pupil response but caused less release of dilation after the end of the sentences. A lower reading span score was associated with higher peak pupil dilation but was not associated with peak latency, release of dilation, or speech recognition in noise. CONCLUSIONS In CI users, speech perception is effortful, even at higher speech recognition scores and high SNRs, indicating that CI users are in a chronic state of increased effort in communication situations. The application of a clinically used NRA did not improve speech perception, nor did it reduce listening effort. Participants with a relatively low WMC exerted relatively more listening effort but did not have better speech reception thresholds in noise.
Collapse
Affiliation(s)
- Gertjan Dingemanse
- Department of Otorhinolaryngology, Head and Neck Surgery, Erasmus University Medical Center, Rotterdam, the Netherlands
| | - André Goedegebure
- Department of Otorhinolaryngology, Head and Neck Surgery, Erasmus University Medical Center, Rotterdam, the Netherlands
| |
Collapse
|
11
|
Wu M, Christiansen S, Fereczkowski M, Neher T. Revisiting Auditory Profiling: Can Cognitive Factors Improve the Prediction of Aided Speech-in-Noise Outcome? Trends Hear 2022; 26:23312165221113889. [PMID: 35942807 PMCID: PMC9373127 DOI: 10.1177/23312165221113889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022] Open
Abstract
Hearing aids (HA) are the most common type of rehabilitation treatment for
age-related hearing loss. However, HA users often obtain limited benefit from
their devices, particularly in noisy environments, and thus many HA candidates
do not use them at all. A possible reason for this could be that current HA
fittings are audiogram-based, that is, they neglect supra-threshold factors. In
an earlier study, an auditory-profiling method was proposed as a basis for more
personalized HA fittings. This method classifies HA users into four profiles
that differ in terms of hearing sensitivity and supra-threshold hearing
abilities. Previously, HA users belonging to these profiles showed significant
differences in terms of speech recognition in noise but not subjective
assessments of speech-in-noise (SIN) outcome. Moreover, large individual
differences within some profiles were observed. The current study therefore
explored if cognitive factors can help explain these differences and improve
aided outcome prediction. Thirty-nine older HA users completed sets of auditory
and SIN tests as well as two tablet-based cognitive measures (the Corsi
block-tapping and trail-making tests). Principal component analyses were applied
to extract the dominant sources of variance both within individual tests
producing many variables and within the three types of tests. Multiple linear
regression analyses performed on the extracted components showed that auditory
factors were related to aided speech recognition in noise but not to subjective
SIN outcome. Cognitive factors were unrelated to aided SIN outcome. Overall,
these findings provide limited support for adding those two cognitive tests to
the profiling of HA users.
Collapse
Affiliation(s)
- Mengfan Wu
- Institute of Clinical Research, Faculty of Health Sciences, 6174University of Southern Denmark, Odense, Denmark.,Research Unit for ORL - Head & Neck Surgery and Audiology, 11286Odense University Hospital & University of Southern Denmark, Odense, Denmark
| | - Stine Christiansen
- Institute of Clinical Research, Faculty of Health Sciences, 6174University of Southern Denmark, Odense, Denmark.,Research Unit for ORL - Head & Neck Surgery and Audiology, 11286Odense University Hospital & University of Southern Denmark, Odense, Denmark
| | - Michal Fereczkowski
- Institute of Clinical Research, Faculty of Health Sciences, 6174University of Southern Denmark, Odense, Denmark.,Research Unit for ORL - Head & Neck Surgery and Audiology, 11286Odense University Hospital & University of Southern Denmark, Odense, Denmark
| | - Tobias Neher
- Institute of Clinical Research, Faculty of Health Sciences, 6174University of Southern Denmark, Odense, Denmark.,Research Unit for ORL - Head & Neck Surgery and Audiology, 11286Odense University Hospital & University of Southern Denmark, Odense, Denmark
| |
Collapse
|
12
|
Beatty-Martínez AL, Guzzardo Tamargo RE, Dussias PE. Phasic pupillary responses reveal differential engagement of attentional control in bilingual spoken language processing. Sci Rep 2021; 11:23474. [PMID: 34873258 PMCID: PMC8648769 DOI: 10.1038/s41598-021-03008-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2021] [Accepted: 11/25/2021] [Indexed: 11/21/2022] Open
Abstract
Language processing is cognitively demanding, requiring attentional resources to efficiently select and extract linguistic information as utterances unfold. Previous research has associated changes in pupil size with increased attentional effort. However, it is unknown whether the behavioral ecology of speakers may differentially affect engagement of attentional resources involved in conversation. For bilinguals, such an act potentially involves competing signals in more than one language and how this competition arises may differ across communicative contexts. We examined changes in pupil size during the comprehension of unilingual and codeswitched speech in a richly-characterized bilingual sample. In a visual-world task, participants saw pairs of objects as they heard instructions to select a target image. Instructions were either unilingual or codeswitched from one language to the other. We found that only bilinguals who use each of their languages in separate communicative contexts and who have high attention ability, show differential attention to unilingual and codeswitched speech. Bilinguals for whom codeswitching is common practice process unilingual and codeswitched speech similarly, regardless of attentional skill. Taken together, these results suggest that bilinguals recruit different language control strategies for distinct communicative purposes. The interactional context of language use critically determines attentional control engagement during language processing.
Collapse
Affiliation(s)
| | - Rosa E Guzzardo Tamargo
- Department of Hispanic Studies, University of Puerto Rico-Río Piedras, San Juan, PR, 00931, USA
| | - Paola E Dussias
- Department of Spanish, Italian and Portuguese, Pennsylvania State University, University Park, PA, 16802, USA
| |
Collapse
|
13
|
Visentin C, Valzolgher C, Pellegatti M, Potente P, Pavani F, Prodi N. A comparison of simultaneously-obtained measures of listening effort: pupil dilation, verbal response time and self-rating. Int J Audiol 2021; 61:561-573. [PMID: 34634214 DOI: 10.1080/14992027.2021.1921290] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
Abstract
OBJECTIVE The aim of this study was to assess to what extent simultaneously-obtained measures of listening effort (task-evoked pupil dilation, verbal response time [RT], and self-rating) could be sensitive to auditory and cognitive manipulations in a speech perception task. The study also aimed to explore the possible relationship between RT and pupil dilation. DESIGN A within-group design was adopted. All participants were administered the Matrix Sentence Test in 12 conditions (signal-to-noise ratios [SNR] of -3, -6, -9 dB; attentional resources focussed vs divided; spatial priors present vs absent). STUDY SAMPLE Twenty-four normal-hearing adults, 20-41 years old (M = 23.5), were recruited in the study. RESULTS A significant effect of the SNR was found for all measures. However, pupil dilation discriminated only partially between the SNRs. Neither of the cognitive manipulations were effective in modulating the measures. No relationship emerged between pupil dilation, RT and self-ratings. CONCLUSIONS RT, pupil dilation, and self-ratings can be obtained simultaneously when administering speech perception tasks, even though some limitations remain related to the absence of a retention period after the listening phase. The sensitivity of the three measures to changes in the auditory environment differs. RTs and self-ratings proved most sensitive to changes in SNR.
Collapse
Affiliation(s)
- Chiara Visentin
- Department of Engineering, University of Ferrara, Ferrara, Italy
| | - Chiara Valzolgher
- Center for Mind/Brain Sciences (CIMeC), University of Trento, Trento, Italy.,Centre de Recherche en Neuroscience de Lyon (CRNL), Integrative, Multisensory, Perception, Action and Cognition Team (IMPACT), Lyon, France
| | | | - Paola Potente
- Center for Mind/Brain Sciences (CIMeC), University of Trento, Trento, Italy
| | - Francesco Pavani
- Center for Mind/Brain Sciences (CIMeC), University of Trento, Trento, Italy.,Centre de Recherche en Neuroscience de Lyon (CRNL), Integrative, Multisensory, Perception, Action and Cognition Team (IMPACT), Lyon, France.,Department of Psychology and Cognitive Sciences (DiPSCo), University of Trento, Trento, Italy
| | - Nicola Prodi
- Department of Engineering, University of Ferrara, Ferrara, Italy
| |
Collapse
|
14
|
Pupillometry reveals cognitive demands of lexical competition during spoken word recognition in young and older adults. Psychon Bull Rev 2021; 29:268-280. [PMID: 34405386 DOI: 10.3758/s13423-021-01991-0] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/27/2021] [Indexed: 12/27/2022]
Abstract
In most contemporary activation-competition frameworks for spoken word recognition, candidate words compete against phonological "neighbors" with similar acoustic properties (e.g., "cap" vs. "cat"). Thus, recognizing words with more competitors should come at a greater cognitive cost relative to recognizing words with fewer competitors, due to increased demands for selecting the correct item and inhibiting incorrect candidates. Importantly, these processes should operate even in the absence of differences in accuracy. In the present study, we tested this proposal by examining differences in processing costs associated with neighborhood density for highly intelligible items presented in quiet. A second goal was to examine whether the cognitive demands associated with increased neighborhood density were greater for older adults compared with young adults. Using pupillometry as an index of cognitive processing load, we compared the cognitive demands associated with spoken word recognition for words with many or fewer neighbors, presented in quiet, for young (n = 67) and older (n = 69) adult listeners. Growth curve analysis of the pupil data indicated that older adults showed a greater evoked pupil response for spoken words than did young adults, consistent with increased cognitive load during spoken word recognition. Words from dense neighborhoods were marginally more demanding to process than words from sparse neighborhoods. There was also an interaction between age and neighborhood density, indicating larger effects of density in young adult listeners. These results highlight the importance of assessing both cognitive demands and accuracy when investigating the mechanisms underlying spoken word recognition.
Collapse
|
15
|
DeRoy Milvae K, Kuchinsky SE, Stakhovskaya OA, Goupell MJ. Dichotic listening performance and effort as a function of spectral resolution and interaural symmetry. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 150:920. [PMID: 34470337 PMCID: PMC8346288 DOI: 10.1121/10.0005653] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/04/2020] [Revised: 06/30/2021] [Accepted: 06/30/2021] [Indexed: 06/13/2023]
Abstract
One potential benefit of bilateral cochlear implants is reduced listening effort in speech-on-speech masking situations. However, the symmetry of the input across ears, possibly related to spectral resolution, could impact binaural benefits. Fifteen young adults with normal hearing performed digit recall with target and interfering digits presented to separate ears and attention directed to the target ear. Recall accuracy and pupil size over time (used as an index of listening effort) were measured for unprocessed, 16-channel vocoded, and 4-channel vocoded digits. Recall accuracy was significantly lower for dichotic (with interfering digits) than for monotic listening. Dichotic recall accuracy was highest when the target was less degraded and the interferer was more degraded. With matched target and interferer spectral resolution, pupil dilation was lower with more degradation. Pupil dilation grew more shallowly over time when the interferer had more degradation. Overall, interferer spectral resolution more strongly affected listening effort than target spectral resolution. These results suggest that interfering speech both lowers performance and increases listening effort, and that the relative spectral resolution of target and interferer affect the listening experience. Ignoring a clearer interferer is more effortful.
Collapse
Affiliation(s)
- Kristina DeRoy Milvae
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Stefanie E Kuchinsky
- Audiology and Speech Pathology Center, Walter Reed National Military Medical Center, Bethesda, Maryland 20889, USA
| | - Olga A Stakhovskaya
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| | - Matthew J Goupell
- Department of Hearing and Speech Sciences, University of Maryland, College Park, Maryland 20742, USA
| |
Collapse
|
16
|
Abstract
Supplemental Digital Content is available in the text. Cognitive load (CL) impairs listeners’ ability to comprehend sentences, recognize words, and identify speech sounds. Recent findings suggest that this effect originates in a disruption of low-level perception of acoustic details. Here, we attempted to quantify such a disruption by measuring the effect of CL (a two-back task) on pure-tone audiometry (PTA) thresholds. We also asked whether the effect of CL on PTA was greater in older adults, on account of their reduced ability to divide cognitive resources between simultaneous tasks. To specify the mechanisms and representations underlying the interface between auditory and cognitive processes, we contrasted CL requiring visual encoding with CL requiring auditory encoding. Finally, the link between the cost of performing PTA under CL, working memory, and speech-in-noise (SiN) perception was investigated and compared between younger and older participants.
Collapse
|
17
|
Lunner T, Alickovic E, Graversen C, Ng EHN, Wendt D, Keidser G. Three New Outcome Measures That Tap Into Cognitive Processes Required for Real-Life Communication. Ear Hear 2021; 41 Suppl 1:39S-47S. [PMID: 33105258 PMCID: PMC7676869 DOI: 10.1097/aud.0000000000000941] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2020] [Accepted: 07/11/2020] [Indexed: 11/29/2022]
Abstract
To increase the ecological validity of outcomes from laboratory evaluations of hearing and hearing devices, it is desirable to introduce more realistic outcome measures in the laboratory. This article presents and discusses three outcome measures that have been designed to go beyond traditional speech-in-noise measures to better reflect realistic everyday challenges. The outcome measures reviewed are: the Sentence-final Word Identification and Recall (SWIR) test that measures working memory performance while listening to speech in noise at ceiling performance; a neural tracking method that produces a quantitative measure of selective speech attention in noise; and pupillometry that measures changes in pupil dilation to assess listening effort while listening to speech in noise. According to evaluation data, the SWIR test provides a sensitive measure in situations where speech perception performance might be unaffected. Similarly, pupil dilation has also shown sensitivity in situations where traditional speech-in-noise measures are insensitive. Changes in working memory capacity and effort mobilization were found at positive signal-to-noise ratios (SNR), that is, at SNRs that might reflect everyday situations. Using stimulus reconstruction, it has been demonstrated that neural tracking is a robust method at determining to what degree a listener is attending to a specific talker in a typical cocktail party situation. Using both established and commercially available noise reduction schemes, data have further shown that all three measures are sensitive to variation in SNR. In summary, the new outcome measures seem suitable for testing hearing and hearing devices under more realistic and demanding everyday conditions than traditional speech-in-noise tests.
Collapse
Affiliation(s)
- Thomas Lunner
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, Linköping University, Linköping, Sweden
- Department of Electrical Engineering, Division Automatic Control, Linköping University, Linköping, Sweden
- Department of Health Technology, Hearing Systems, Technical University of Denmark, Lyngby, Denmark
| | - Emina Alickovic
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- Department of Electrical Engineering, Division Automatic Control, Linköping University, Linköping, Sweden
| | | | - Elaine Hoi Ning Ng
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, Linköping University, Linköping, Sweden
- Oticon A/S, Kongebakken, Denmark
| | - Dorothea Wendt
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- Department of Health Technology, Hearing Systems, Technical University of Denmark, Lyngby, Denmark
| | - Gitte Keidser
- Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, Linköping University, Linköping, Sweden
| |
Collapse
|
18
|
Brännström KJ, Rudner M, Carlie J, Sahlén B, Gulz A, Andersson K, Johansson R. Listening effort and fatigue in native and non-native primary school children. J Exp Child Psychol 2021; 210:105203. [PMID: 34118494 DOI: 10.1016/j.jecp.2021.105203] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2021] [Revised: 04/15/2021] [Accepted: 05/14/2021] [Indexed: 11/26/2022]
Abstract
Background noise makes listening effortful and may lead to fatigue. This may compromise classroom learning, especially for children with a non-native background. In the current study, we used pupillometry to investigate listening effort and fatigue during listening comprehension under typical (0 dB signal-to-noise ratio [SNR]) and favorable (+10 dB SNR) listening conditions in 63 Swedish primary school children (7-9 years of age) performing a narrative speech-picture verification task. Our sample comprised both native (n = 25) and non-native (n = 38) speakers of Swedish. Results revealed greater pupil dilation, indicating more listening effort, in the typical listening condition compared with the favorable listening condition, and it was primarily the non-native speakers who contributed to this effect (and who also had lower performance accuracy than the native speakers). Furthermore, the native speakers had greater pupil dilation during successful trials, whereas the non-native speakers showed greatest pupil dilation during unsuccessful trials, especially in the typical listening condition. This set of results indicates that whereas native speakers can apply listening effort to good effect, non-native speakers may have reached their effort ceiling, resulting in poorer listening comprehension. Finally, we found that baseline pupil size decreased over trials, which potentially indicates more listening-related fatigue, and this effect was greater in the typical listening condition compared with the favorable listening condition. Collectively, these results provide novel insight into the underlying dynamics of listening effort, fatigue, and listening comprehension in typical classroom conditions compared with favorable classroom conditions, and they demonstrate for the first time how sensitive this interplay is to language experience.
Collapse
Affiliation(s)
- K Jonas Brännström
- Logopedics, Phoniatrics and Audiology, Department of Clinical Sciences in Lund, Lund University, 221 85 Lund, Sweden
| | - Mary Rudner
- Linnaeus Centre HEAD, Department of Behavioural Sciences and Learning, Linköping University, 581 83 Linköping, Sweden
| | - Johanna Carlie
- Logopedics, Phoniatrics and Audiology, Department of Clinical Sciences in Lund, Lund University, 221 85 Lund, Sweden
| | - Birgitta Sahlén
- Logopedics, Phoniatrics and Audiology, Department of Clinical Sciences in Lund, Lund University, 221 85 Lund, Sweden
| | - Agneta Gulz
- Division of Cognitive Science, Lund University, 221 00 Lund, Sweden
| | - Ketty Andersson
- Logopedics, Phoniatrics and Audiology, Department of Clinical Sciences in Lund, Lund University, 221 85 Lund, Sweden
| | - Roger Johansson
- Department of Psychology, Lund University, 221 00 Lund, Sweden.
| |
Collapse
|
19
|
The Effects of Task Difficulty Predictability and Noise Reduction on Recall Performance and Pupil Dilation Responses. Ear Hear 2021; 42:1668-1679. [PMID: 33859121 PMCID: PMC8542077 DOI: 10.1097/aud.0000000000001053] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Communication requires cognitive processes which are not captured by traditional speech understanding tests. Under challenging listening situations, more working memory resources are needed to process speech, leaving fewer resources available for storage. The aim of the current study was to investigate the effect of task difficulty predictability, that is, knowing versus not knowing task difficulty in advance, and the effect of noise reduction on working memory resource allocation to processing and storage of speech heard in background noise. For this purpose, an "offline" behavioral measure, the Sentence-Final Word Identification and Recall (SWIR) test, and an "online" physiological measure, pupillometry, were combined. Moreover, the outcomes of the two measures were compared to investigate whether they reflect the same processes related to resource allocation. DESIGN Twenty-four experienced hearing aid users with moderate to moderately severe hearing loss participated in this study. The SWIR test and pupillometry were measured simultaneously with noise reduction in the test hearing aids activated and deactivated in a background noise composed of four-talker babble. The task of the SWIR test is to listen to lists of sentences, repeat the last word immediately after each sentence and recall the repeated words when the list is finished. The sentence baseline dilation, which is defined as the mean pupil dilation before each sentence, and task-evoked peak pupil dilation (PPD) were analyzed over the course of the lists. The task difficulty predictability was manipulated by including lists of three, five, and seven sentences. The test was conducted over two sessions, one during which the participants were informed about list length before each list (predictable task difficulty) and one during which they were not (unpredictable task difficulty). RESULTS The sentence baseline dilation was higher when task difficulty was unpredictable compared to predictable, except at the start of the list, where there was no difference. The PPD tended to be higher at the beginning of the list, this pattern being more prominent when task difficulty was unpredictable. Recall performance was better and sentence baseline dilation was higher when noise reduction was on, especially toward the end of longer lists. There was no effect of noise reduction on PPD. CONCLUSIONS Task difficulty predictability did not have an effect on resource allocation, since recall performance was similar independently of whether task difficulty was predictable or unpredictable. The higher sentence baseline dilation when task difficulty was unpredictable likely reflected a difference in the recall strategy or higher degree of task engagement/alertness or arousal. Hence, pupillometry captured processes which the SWIR test does not capture. Noise reduction frees up resources to be used for storage of speech, which was reflected in the better recall performance and larger sentence baseline dilation toward the end of the list when noise reduction was on. Thus, both measures captured different temporal aspects of the same processes related to resource allocation with noise reduction on and off.
Collapse
|
20
|
Zhang Y, Lehmann A, Deroche M. Disentangling listening effort and memory load beyond behavioural evidence: Pupillary response to listening effort during a concurrent memory task. PLoS One 2021; 16:e0233251. [PMID: 33657100 PMCID: PMC7928507 DOI: 10.1371/journal.pone.0233251] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2020] [Accepted: 02/15/2021] [Indexed: 11/18/2022] Open
Abstract
Recent research has demonstrated that pupillometry is a robust measure for quantifying listening effort. However, pupillary responses in listening situations where multiple cognitive functions are engaged and sustained over a period of time remain hard to interpret. This limits our conceptualisation and understanding of listening effort in realistic situations, because rarely in everyday life are people challenged by one task at a time. Therefore, the purpose of this experiment was to reveal the dynamics of listening effort in a sustained listening condition using a word repeat and recall task. Words were presented in quiet and speech-shaped noise at different signal-to-noise ratios (SNR): 0dB, 7dB, 14dB and quiet. Participants were presented with lists of 10 words, and required to repeat each word after its presentation. At the end of the list, participants either recalled as many words as possible or moved on to the next list. Simultaneously, their pupil dilation was recorded throughout the whole experiment. When only word repeating was required, peak pupil dilation (PPD) was bigger in 0dB versus other conditions; whereas when recall was required, PPD showed no difference among SNR levels and PPD in 0dB was smaller than repeat-only condition. Baseline pupil diameter and PPD followed different variation patterns across the 10 serial positions within a block for conditions requiring recall: baseline pupil diameter built up progressively and plateaued in the later positions (but shot up when listeners were recalling the previously heard words from memory); PPD decreased at a pace quicker than in repeat-only condition. The current findings demonstrate that additional cognitive load during a speech intelligibility task could disturb the well-established relation between pupillary response and listening effort. Both the magnitude and temporal pattern of task-evoked pupillary response differ greatly in complex listening conditions, urging for more listening effort studies in complex and realistic listening situations.
Collapse
Affiliation(s)
- Yue Zhang
- Department of Otolaryngology, McGill University, Montreal, Canada
- Centre for Research on Brain, Language and Music, Montreal, Canada
- Laboratory for Brain, Music and Sound Research, Montreal, Canada
- Centre for Interdisciplinary Research in Music Media and Technology, Montreal, Canada
- * E-mail:
| | - Alexandre Lehmann
- Department of Otolaryngology, McGill University, Montreal, Canada
- Centre for Research on Brain, Language and Music, Montreal, Canada
- Laboratory for Brain, Music and Sound Research, Montreal, Canada
- Centre for Interdisciplinary Research in Music Media and Technology, Montreal, Canada
| | - Mickael Deroche
- Department of Otolaryngology, McGill University, Montreal, Canada
- Centre for Research on Brain, Language and Music, Montreal, Canada
- Laboratory for Brain, Music and Sound Research, Montreal, Canada
- Centre for Interdisciplinary Research in Music Media and Technology, Montreal, Canada
- Department of Psychology, Concordia University, Montreal, Canada
| |
Collapse
|
21
|
McLaughlin DJ, Braver TS, Peelle JE. Measuring the Subjective Cost of Listening Effort Using a Discounting Task. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:337-347. [PMID: 33439751 PMCID: PMC8632478 DOI: 10.1044/2020_jslhr-20-00086] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Purpose Objective measures of listening effort have been gaining prominence, as they provide metrics to quantify the difficulty of understanding speech under a variety of circumstances. A key challenge has been to develop paradigms that enable the complementary measurement of subjective listening effort in a quantitatively precise manner. In this study, we introduce a novel decision-making paradigm to examine age-related and individual differences in subjective effort during listening. Method Older and younger adults were presented with spoken sentences mixed with speech-shaped noise at multiple signal-to-noise ratios (SNRs). On each trial, subjects were offered the choice between completing an easier listening trial (presented at +20 dB SNR) for a smaller monetary reward and completing a harder listening trial (presented at either +4, 0, -4, -8, or -12 dB SNR) for a greater monetary reward. By varying the amount of the reward offered for the easier option, the subjective value of performing effortful listening trials at each SNR could be assessed. Results Older adults discounted the value of effortful listening to a greater degree than young adults, opting to accept less money in order to avoid more difficult SNRs. Additionally, older adults with poorer hearing and smaller working memory capacities were more likely to choose easier trials; however, in younger adults, no relationship with hearing or working memory was found. Self-reported measures of economic status did not affect these relationships. Conclusions These findings suggest that subjective listening effort depends on factors including, but not necessarily limited to, hearing and working memory. Additionally, this study demonstrates that economic decision-making paradigms can be a useful approach for assessing subjective listening effort and may prove beneficial in future research.
Collapse
Affiliation(s)
- Drew J. McLaughlin
- Department of Psychological & Brain Sciences, Washington University in St. Louis, MO
| | - Todd S. Braver
- Department of Psychological & Brain Sciences, Washington University in St. Louis, MO
| | | |
Collapse
|
22
|
Książek P, Zekveld AA, Wendt D, Fiedler L, Lunner T, Kramer SE. Effect of Speech-to-Noise Ratio and Luminance on a Range of Current and Potential Pupil Response Measures to Assess Listening Effort. Trends Hear 2021; 25:23312165211009351. [PMID: 33926329 PMCID: PMC8111552 DOI: 10.1177/23312165211009351] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2020] [Revised: 02/11/2021] [Accepted: 03/05/2021] [Indexed: 11/17/2022] Open
Abstract
In hearing research, pupillometry is an established method of studying listening effort. The focus of this study was to evaluate several pupil measures extracted from the Task-Evoked Pupil Responses (TEPRs) in speech-in-noise test. A range of analysis approaches was applied to extract these pupil measures, namely (a) pupil peak dilation (PPD); (b) mean pupil dilation (MPD); (c) index of pupillary activity; (d) growth curve analysis (GCA); and (e) principal component analysis (PCA). The effect of signal-to-noise ratio (SNR; Data Set A: -20 dB, -10 dB, +5 dB SNR) and luminance (Data Set B: 0.1 cd/m2, 360 cd/m2) on the TEPRs were investigated. Data Sets A and B were recorded during a speech-in-noise test and included TEPRs from 33 and 27 normal-hearing native Dutch speakers, respectively. The main results were as follows: (a) A significant effect of SNR was revealed for all pupil measures extracted in the time domain (PPD, MPD, GCA, PCA); (b) Two time series analysis approaches (GCA, PCA) provided modeled temporal profiles of TEPRs (GCA); and time windows spanning subtasks performed in a speech-in-noise test (PCA); and (c) All pupil measures revealed a significant effect of luminance. In conclusion, multiple pupil measures showed similar effects of SNR, suggesting that effort may be reflected in multiple aspects of TEPR. Moreover, a direct analysis of the pupil time course seems to provide a more holistic view of TEPRs, yet further research is needed to understand and interpret its measures. Further research is also required to find pupil measures less sensitive to changes in luminance.
Collapse
Affiliation(s)
- Patrycja Książek
- Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology-Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, the Netherlands
- Eriksholm Research Centre, Snekkersten, Denmark
| | - Adriana A. Zekveld
- Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology-Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, the Netherlands
| | - Dorothea Wendt
- Eriksholm Research Centre, Snekkersten, Denmark
- Department of Health Technology, Technical University of Denmark, Lyngby, Denmark
| | | | | | - Sophia E. Kramer
- Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology-Head and Neck Surgery, Ear and Hearing, Amsterdam Public Health Research Institute, Amsterdam, the Netherlands
| |
Collapse
|
23
|
Investigating the Influences of Task Demand and Reward on Cardiac Pre-Ejection Period Reactivity During a Speech-in-Noise Task. Ear Hear 2020; 42:718-731. [PMID: 33201048 PMCID: PMC8088822 DOI: 10.1097/aud.0000000000000971] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
OBJECTIVES Effort investment during listening varies as a function of task demand and motivation. Several studies have manipulated both these factors to elicit and measure changes in effort associated with listening. The cardiac pre-ejection period (PEP) is a relatively novel measure in the field of cognitive hearing science. This measure, which reflects sympathetic nervous system activity on the heart, has previously been implemented during a tone discrimination task but not during a speech-in-noise task. Therefore, the primary goal of this study was to explore the influences of signal to noise ratio (SNR) and monetary reward level on PEP reactivity during a speech-in-noise task. DESIGN Thirty-two participants with normal hearing (mean age = 22.22 years, SD = 3.03) were recruited at VU University Medical Center. Participants completed a Dutch speech-in-noise test with a single-interfering-talker masking noise. Six fixed SNRs, selected to span the entire psychometric performance curve, were presented in a block-wise fashion. Participants could earn a low (€0.20) or high (€5.00) reward by obtaining a score of ≥70% of words correct in each block. The authors analyzed PEP reactivity: the change in PEP measured during the task, relative to the baseline during rest. Two separate methods of PEP analysis were used, one including data from the whole task block and the other including data obtained during presentation of the target sentences only. After each block, participants rated their effort investment, performance, tendency to give up, and the perceived difficulty of the task. They also completed the need for recovery questionnaire and the reading span test, which are indices of additional factors (fatigue and working memory capacity, respectively) that are known to influence listening effort. RESULTS Average sentence perception scores ranged from 2.73 to 91.62%, revealing a significant effect of SNR. In addition, an improvement in performance was elicited by the high, compared to the low reward level. A linear relationship between SNR and PEP reactivity was demonstrated: at the lower SNRs PEP reactivity was the most negative, indicating greater effort investment compared to the higher SNRs. The target stimuli method of PEP analysis was more sensitive to this effect than the block-wise method. Contrary to expectations, no significant impact of reward on PEP reactivity was found in the present dataset. Also, there was no physiological evidence that participants were disengaged, even when performance was poor. A significant correlation between need for recovery scores and average PEP reactivity was demonstrated, indicating that a lower need for recovery was associated with less effort investment. CONCLUSIONS This study successfully implemented the measurement of PEP during a standard speech-in-noise test and included two distinct methods of PEP analysis. The results revealed for the first time that PEP reactivity varies linearly with task demand during a speech-in-noise task, although the effect size was small. No effect of reward on PEP was demonstrated. Finally, participants with a higher need for recovery score invested more effort, as shown by average PEP reactivity, than those with a lower need for recovery score.
Collapse
|
24
|
Kim S, Choi I, Schwalje AT, Kim K, Lee JH. Auditory Working Memory Explains Variance in Speech Recognition in Older Listeners Under Adverse Listening Conditions. Clin Interv Aging 2020; 15:395-406. [PMID: 32231429 PMCID: PMC7085334 DOI: 10.2147/cia.s241976] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2019] [Accepted: 02/13/2020] [Indexed: 12/27/2022] Open
Abstract
Introduction Older listeners have difficulty understanding speech in unfavorable listening conditions. To compensate for acoustic degradation, cognitive processing skills, such as working memory, need to be engaged. Despite prior findings on the association between working memory and speech recognition in various listening conditions, it is not yet clear whether the modality of stimuli presentation for working memory tasks should be auditory or visual. Given the modality-specific characteristics of working memory, we hypothesized that auditory working memory capacity could predict speech recognition performance in adverse listening conditions for older listeners and that the contribution of auditory working memory to speech recognition would depend on the task and listening condition. Methods Seventy-six older listeners and twenty younger listeners completed four kinds of auditory working memory tasks, including digit and speech span tasks, and sentence recognition tasks in four different listening conditions having multi-talker noise and time-compression. For older listeners, cognitive function was screened using the Mini-Mental Status Examination, and audibility was assured. Results Auditory working memory, as measured by listening span, significantly predicted speech recognition performance in adverse listening conditions for older listeners. A linear regression model showed speech recognition performance for older listeners could be explained by auditory working memory whilst controlling for the impact of age and hearing sensitivity. Discussion Measuring working memory in the auditory modality facilitated explaining the variance in speech recognition in adverse listening conditions for older listeners. The linguistic features and the complexity of the auditory stimuli may affect the association between working memory and speech recognition performance. Conclusion We demonstrated the contribution of auditory working memory to speech recognition in unfavorable listening conditions in older populations. Taking the modality-specific characteristics of working memory into account may be a key to better understand the difficulty in speech recognition in daily listening conditions for older listeners.
Collapse
Affiliation(s)
- Subong Kim
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA 52242, USA
| | - Inyong Choi
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA 52242, USA.,Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA
| | - Adam T Schwalje
- Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA
| | - KyooSang Kim
- Department of Occupational Environmental Medicine, Seoul Medical Centerr, Seoul 02053, South Korea
| | - Jae Hee Lee
- Department of Audiology and Speech-Language Pathology, HUGS Center for Hearing and Speech Research, Hallym University of Graduate Studies, Seoul 06197, South Korea
| |
Collapse
|
25
|
Salorio-Corbetto M, Baer T, Stone MA, Moore BCJ. Effect of the number of amplitude-compression channels and compression speed on speech recognition by listeners with mild to moderate sensorineural hearing loss. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2020; 147:1344. [PMID: 32237835 DOI: 10.1121/10.0000804] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/08/2019] [Accepted: 02/09/2020] [Indexed: 06/11/2023]
Abstract
The use of a large number of amplitude-compression channels in hearing aids has potential advantages, such as the ability to compensate for variations in loudness recruitment across frequency and provide appropriate frequency-response shaping. However, sound quality and speech intelligibility could be adversely affected due to reduction of spectro-temporal contrast and distortion, especially when fast-acting compression is used. This study assessed the effect of the number of channels and compression speed on speech recognition when the multichannel processing was used solely to implement amplitude compression, and not for frequency-response shaping. Computer-simulated hearing aids were used. The frequency-dependent insertion gains for speech with a level of 65 dB sound pressure level were applied using a single filter before the signal was filtered into compression channels. Fast-acting (attack, 10 ms; release, 100 ms) or slow-acting (attack, 50 ms; release, 3000 ms) compression using 3, 6, 12, and 22 channels was applied subsequently. Using a sentence recognition task with speech in two- and eight-talker babble at three different signal-to-babble ratios (SBRs), 20 adults with sensorineural hearing loss were tested. The number of channels and compression speed had no significant effect on speech recognition, regardless of babble type or SBR.
Collapse
Affiliation(s)
- Marina Salorio-Corbetto
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom
| | - Thomas Baer
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom
| | - Michael A Stone
- Division of Human Communication, Development and Hearing, University of Manchester, Oxford Road, Manchester M13 9PL, United Kingdom
| | - Brian C J Moore
- Department of Experimental Psychology, University of Cambridge, Downing Street, Cambridge CB2 3EB, United Kingdom
| |
Collapse
|
26
|
Miller MK, Calandruccio L, Buss E, McCreery RW, Oleson J, Rodriguez B, Leibold LJ. Masked English Speech Recognition Performance in Younger and Older Spanish-English Bilingual and English Monolingual Children. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:4578-4591. [PMID: 31830845 PMCID: PMC7839054 DOI: 10.1044/2019_jslhr-19-00059] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/09/2019] [Accepted: 05/09/2019] [Indexed: 06/01/2023]
Abstract
Purpose The purpose of this study was to compare masked English speech recognition thresholds between Spanish-English bilingual and English monolingual children and to evaluate effects of age, maternal education, and English receptive language abilities on individual differences in masked speech recognition. Method Forty-three Spanish-English bilingual children and 42 English monolingual children completed an English sentence recognition task in 2 masker conditions: (a) speech-shaped noise and (b) 2-talker English speech. Two age groups of children, younger (5-6 years) and older (9-10 years), were tested. The predictors of masked speech recognition performance were evaluated using 2 mixed-effects regression models. In the 1st model, fixed effects were age group (younger children vs. older children), language group (bilingual vs. monolingual), and masker type (speech-shaped noise vs. 2-talker speech). In the 2nd model, the fixed effects of receptive English vocabulary scores and maternal education level were also included. Results Younger children performed more poorly than older children, but no significant difference in masked speech recognition was observed between bilingual and monolingual children for either age group when English proficiency and maternal education were also included in the model. English language abilities fell within age-appropriate norms for both groups, but individual children with larger receptive vocabularies in English tended to show better recognition; this effect was stronger for younger children than for older children. Speech reception thresholds for all children were lower in the speech-shaped noise masker than in the 2-talker speech masker. Conclusions Regardless of age, similar masked speech recognition was observed for Spanish-English bilingual and English monolingual children tested in this study when receptive English language abilities were accounted for. Receptive English vocabulary scores were associated with better masked speech recognition performance for both bilinguals and monolinguals, with a stronger relationship observed for younger children than older children. Further investigation involving a Spanish-dominant bilingual sample is warranted given the high English language proficiency of children included in this study.
Collapse
Affiliation(s)
- Margaret K. Miller
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| | - Lauren Calandruccio
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, OH
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, University of North Carolina at Chapel Hill
| | - Ryan W. McCreery
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| | - Jacob Oleson
- Department of Biostatistics, University of Iowa, Iowa City
| | - Barbara Rodriguez
- Department of Speech and Hearing Sciences, The University of New Mexico, Albuquerque
| | - Lori J. Leibold
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, NE
| |
Collapse
|
27
|
Abstract
It is widely accepted that seeing a talker improves a listener's ability to understand what a talker is saying in background noise (e.g., Erber, 1969; Sumby & Pollack, 1954). The literature is mixed, however, regarding the influence of the visual modality on the listening effort required to recognize speech (e.g., Fraser, Gagné, Alepins, & Dubois, 2010; Sommers & Phelps, 2016). Here, we present data showing that even when the visual modality robustly benefits recognition, processing audiovisual speech can still result in greater cognitive load than processing speech in the auditory modality alone. We show using a dual-task paradigm that the costs associated with audiovisual speech processing are more pronounced in easy listening conditions, in which speech can be recognized at high rates in the auditory modality alone-indeed, effort did not differ between audiovisual and audio-only conditions when the background noise was presented at a more difficult level. Further, we show that though these effects replicate with different stimuli and participants, they do not emerge when effort is assessed with a recall paradigm rather than a dual-task paradigm. Together, these results suggest that the widely cited audiovisual recognition benefit may come at a cost under more favorable listening conditions, and add to the growing body of research suggesting that various measures of effort may not be tapping into the same underlying construct (Strand et al., 2018).
Collapse
|
28
|
Moberly AC, Reed J. Making Sense of Sentences: Top-Down Processing of Speech by Adult Cochlear Implant Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:2895-2905. [PMID: 31330118 PMCID: PMC6802905 DOI: 10.1044/2019_jslhr-h-18-0472] [Citation(s) in RCA: 35] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/30/2018] [Revised: 03/20/2019] [Accepted: 04/12/2019] [Indexed: 05/03/2023]
Abstract
Purpose Speech recognition relies upon a listener's successful pairing of the acoustic-phonetic details from the bottom-up input with top-down linguistic processing of the incoming speech stream. When the speech is spectrally degraded, such as through a cochlear implant (CI), this role of top-down processing is poorly understood. This study explored the interactions of top-down processing, specifically the use of semantic context during sentence recognition, and the relative contributions of different neurocognitive functions during speech recognition in adult CI users. Method Data from 41 experienced adult CI users were collected and used in analyses. Participants were tested for recognition and immediate repetition of speech materials in the clear. They were asked to repeat 2 sets of sentence materials, 1 that was semantically meaningful and 1 that was syntactically appropriate but semantically anomalous. Participants also were tested on 4 visual measures of neurocognitive functioning to assess working memory capacity (Digit Span; Wechsler, 2004), speed of lexical access (Test of Word Reading Efficiency; Torgeson, Wagner, & Rashotte, 1999), inhibitory control (Stroop; Stroop, 1935), and nonverbal fluid reasoning (Raven's Progressive Matrices; Raven, 2000). Results Individual listeners' inhibitory control predicted recognition of meaningful sentences when controlling for performance on anomalous sentences, our proxy for the quality of the bottom-up input. Additionally, speed of lexical access and nonverbal reasoning predicted recognition of anomalous sentences. Conclusions Findings from this study identified inhibitory control as a potential mechanism at work when listeners make use of semantic context during sentence recognition. Moreover, speed of lexical access and nonverbal reasoning were associated with recognition of sentences that lacked semantic context. These results motivate the development of improved comprehensive rehabilitative approaches for adult patients with CIs to optimize use of top-down processing and underlying core neurocognitive functions.
Collapse
Affiliation(s)
- Aaron C. Moberly
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
| | - Jessa Reed
- Department of Otolaryngology, The Ohio State University Wexner Medical Center, Columbus
| |
Collapse
|
29
|
Francis AL, Love J. Listening effort: Are we measuring cognition or affect, or both? WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2019; 11:e1514. [PMID: 31381275 DOI: 10.1002/wcs.1514] [Citation(s) in RCA: 46] [Impact Index Per Article: 9.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2019] [Revised: 07/07/2019] [Accepted: 07/10/2019] [Indexed: 12/14/2022]
Abstract
Listening effort is increasingly recognized as a factor in communication, particularly for and with nonnative speakers, for the elderly, for individuals with hearing impairment and/or for those working in noise. However, as highlighted by McGarrigle et al., International Journal of Audiology, 2014, 53, 433-445, the term "listening effort" encompasses a wide variety of concepts, including the engagement and control of multiple possibly distinct neural systems for information processing, and the affective response to the expenditure of those resources in a given context. Thus, experimental or clinical methods intended to objectively quantify listening effort may ultimately reflect a complex interaction between the operations of one or more of those information processing systems, and/or the affective and motivational response to the demand on those systems. Here we examine theoretical, behavioral, and psychophysiological factors related to resolving the question of what we are measuring, and why, when we measure "listening effort." This article is categorized under: Linguistics > Language in Mind and Brain Psychology > Theory and Methods Psychology > Attention Psychology > Emotion and Motivation.
Collapse
Affiliation(s)
- Alexander L Francis
- Department of Speech, Language and Hearing Sciences, Purdue University, West Lafayette, Indiana
| | - Jordan Love
- Department of Speech, Language and Hearing Sciences, Purdue University, West Lafayette, Indiana
| |
Collapse
|
30
|
Lau MK, Hicks C, Kroll T, Zupancic S. Effect of Auditory Task Type on Physiological and Subjective Measures of Listening Effort in Individuals With Normal Hearing. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2019; 62:1549-1560. [PMID: 31063438 DOI: 10.1044/2018_jslhr-h-17-0473] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Purpose Listening effort has traditionally been measured using subjective rating scales and behavioral measures. Recent physiological measures of listening effort have utilized pupil dilation. Using a combination of physiological and subjective measures of listening effort, this study aimed to identify differences in listening effort during 2 auditory tasks: sentence recognition and word recognition. Method Pupil dilation and subjective ratings of listening effort were obtained for auditory tasks utilizing AzBio sentences recognition and Northwestern University Auditory Test No. 6 words recognition, across 3 listening situations: in quiet, at +6 dB signal-to-noise ratio, and at 0 dB signal-to-noise ratio. Task accuracy was recorded for each of the 6 conditions, as well as peak pupil dilation and a subjective rating of listening effort. Results A significant impact of listening situation (quiet vs. noise) and task type (sentence recognition vs. word recognition) on both physiological and subjective measures was found. There was a significant interaction between listening situation and task type, suggesting that contextual cues may only be beneficial when audibility is uncompromised. The current study found no correlation between the physiological and subjective measures, possibly suggesting that these measures analyze different aspects of cognitive effort in a listening task.
Collapse
Affiliation(s)
- Marcy K Lau
- Department of Audiology & Speech Pathology, East Tennessee State University, Johnson City
| | - Candace Hicks
- Department of Speech, Language, & Hearing Sciences, Texas Tech University Health Sciences Center, Lubbock
| | - Tobias Kroll
- Department of Speech, Language, & Hearing Sciences, Texas Tech University Health Sciences Center, Lubbock
| | - Steven Zupancic
- Department of Speech, Language, & Hearing Sciences, Texas Tech University Health Sciences Center, Lubbock
| |
Collapse
|
31
|
Peinkhofer C, Knudsen GM, Moretti R, Kondziella D. Cortical modulation of pupillary function: systematic review. PeerJ 2019; 7:e6882. [PMID: 31119083 PMCID: PMC6510220 DOI: 10.7717/peerj.6882] [Citation(s) in RCA: 28] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2019] [Accepted: 03/26/2019] [Indexed: 12/25/2022] Open
Abstract
BACKGROUND The pupillary light reflex is the main mechanism that regulates the pupillary diameter; it is controlled by the autonomic system and mediated by subcortical pathways. In addition, cognitive and emotional processes influence pupillary function due to input from cortical innervation, but the exact circuits remain poorly understood. We performed a systematic review to evaluate the mechanisms behind pupillary changes associated with cognitive efforts and processing of emotions and to investigate the cerebral areas involved in cortical modulation of the pupillary light reflex. METHODOLOGY We searched multiple databases until November 2018 for studies on cortical modulation of pupillary function in humans and non-human primates. Of 8,809 papers screened, 258 studies were included. RESULTS Most investigators focused on pupillary dilatation and/or constriction as an index of cognitive and emotional processing, evaluating how changes in pupillary diameter reflect levels of attention and arousal. Only few tried to correlate specific cerebral areas to pupillary changes, using either cortical activation models (employing micro-stimulation of cortical structures in non-human primates) or cortical lesion models (e.g., investigating patients with stroke and damage to salient cortical and/or subcortical areas). Results suggest the involvement of several cortical regions, including the insular cortex (Brodmann areas 13 and 16), the frontal eye field (Brodmann area 8) and the prefrontal cortex (Brodmann areas 11 and 25), and of subcortical structures such as the locus coeruleus and the superior colliculus. CONCLUSIONS Pupillary dilatation occurs with many kinds of mental or emotional processes, following sympathetic activation or parasympathetic inhibition. Conversely, pupillary constriction may occur with anticipation of a bright stimulus (even in its absence) and relies on a parasympathetic activation. All these reactions are controlled by subcortical and cortical structures that are directly or indirectly connected to the brainstem pupillary innervation system.
Collapse
Affiliation(s)
- Costanza Peinkhofer
- Department of Neurology, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Medical Faculty, University of Trieste, Trieste, Italy
| | - Gitte M. Knudsen
- Department of Neurology, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Neurobiology Research Unit, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Faculty of Health and Medical Science, University of Copenhagen, Copenhagen, Denmark
| | - Rita Moretti
- Medical Faculty, University of Trieste, Trieste, Italy
- Department of Medical, Surgical and Health Sciences, Neurological Unit, Trieste University Hospital, Cattinara, Trieste, Italy
| | - Daniel Kondziella
- Department of Neurology, Rigshospitalet, Copenhagen University Hospital, Copenhagen, Denmark
- Faculty of Health and Medical Science, University of Copenhagen, Copenhagen, Denmark
- Department of Neuroscience, Norwegian University of Technology and Science, Trondheim, Norway
| |
Collapse
|
32
|
Koelewijn T, van Haastrecht JAP, Kramer SE. Pupil Responses of Adults With Traumatic Brain Injury During Processing of Speech in Noise. Trends Hear 2019; 22:2331216518811444. [PMID: 30482105 PMCID: PMC6277755 DOI: 10.1177/2331216518811444] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Previous research has shown the effects of task demands on pupil responses in both normal hearing (NH) and hearing impaired (HI) adults. One consistent finding is that HI listeners have smaller pupil dilations at low levels of speech recognition performance (≤50%). This study aimed to examine the pupil dilation in adults with a normal pure-tone audiogram who experience serious difficulties when processing speech-in-noise. Hence, 20 adults, aged 26 to 62 years, with traumatic brain injury (TBI) or cerebrovascular accident (CVA) but with a normal audiogram participated. Their pupil size was recorded while they listened to sentences masked by fluctuating noise or interfering speech at 50% and 84% intelligibility. In each condition, participants rated their perceived performance, effort, and task persistence. In addition, participants performed the text reception threshold task—a visual sentence completion task—that measured language-related processing. Data were compared with those of age-matched NH and HI participants with no neurological problems obtained in earlier studies using the same setup and design. The TBI group had the same pure-tone audiogram and text reception threshold scores as the NH listeners, yet their speech reception thresholds were significantly worse. Although the pupil dilation responses on average did not differ between groups, self-rated effort scores were highest in the TBI group. Results of a correlation analyses showed that TBI participants with worse speech reception thresholds had a smaller pupil response. We speculate that increased distractibility or fatigue affected the ability of TBI participants to allocate effort during speech perception in noise.
Collapse
Affiliation(s)
- Thomas Koelewijn
- 1 Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology - Head and Neck surgery, Ear & Hearing, Amsterdam Public Health research institute, Amsterdam, The Netherlands
| | - José A P van Haastrecht
- 1 Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology - Head and Neck surgery, Ear & Hearing, Amsterdam Public Health research institute, Amsterdam, The Netherlands
| | - Sophia E Kramer
- 1 Amsterdam UMC, Vrije Universiteit Amsterdam, Otolaryngology - Head and Neck surgery, Ear & Hearing, Amsterdam Public Health research institute, Amsterdam, The Netherlands
| |
Collapse
|
33
|
Müller JA, Wendt D, Kollmeier B, Debener S, Brand T. Effect of Speech Rate on Neural Tracking of Speech. Front Psychol 2019; 10:449. [PMID: 30906273 PMCID: PMC6418035 DOI: 10.3389/fpsyg.2019.00449] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2018] [Accepted: 02/14/2019] [Indexed: 12/03/2022] Open
Abstract
Speech comprehension requires effort in demanding listening situations. Selective attention may be required for focusing on a specific talker in a multi-talker environment, may enhance effort by requiring additional cognitive resources, and is known to enhance the neural representation of the attended talker in the listener's neural response. The aim of the study was to investigate the relation of listening effort, as quantified by subjective effort ratings and pupil dilation, and neural speech tracking during sentence recognition. Task demands were varied using sentences with varying levels of linguistic complexity and using two different speech rates in a picture-matching paradigm with 20 normal-hearing listeners. The participants' task was to match the acoustically presented sentence with a picture presented before the acoustic stimulus. Afterwards they rated their perceived effort on a categorical effort scale. During each trial, pupil dilation (as an indicator of listening effort) and electroencephalogram (as an indicator of neural speech tracking) were recorded. Neither measure was significantly affected by linguistic complexity. However, speech rate showed a strong influence on subjectively rated effort, pupil dilation, and neural tracking. The neural tracking analysis revealed a shorter latency for faster sentences, which may reflect a neural adaptation to the rate of the input. No relation was found between neural tracking and listening effort, even though both measures were clearly influenced by speech rate. This is probably due to factors that influence both measures differently. Consequently, the amount of listening effort is not clearly represented in the neural tracking.
Collapse
Affiliation(s)
- Jana Annina Müller
- Cluster of Excellence ‘Hearing4all’, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
- Medizinische Physik, Department of Medical Physics and Acoustics, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Dorothea Wendt
- Hearing Systems, Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Kongens Lyngby, Denmark
- Eriksholm Research Centre, Snekkersten, Denmark
| | - Birger Kollmeier
- Cluster of Excellence ‘Hearing4all’, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
- Medizinische Physik, Department of Medical Physics and Acoustics, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Stefan Debener
- Cluster of Excellence ‘Hearing4all’, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
- Neuropsychology Lab, Department of Psychology, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
| | - Thomas Brand
- Cluster of Excellence ‘Hearing4all’, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
- Medizinische Physik, Department of Medical Physics and Acoustics, Carl von Ossietzky Universität Oldenburg, Oldenburg, Germany
| |
Collapse
|
34
|
Lee SJ, Kim H, Kim LS, Kim JH, Park KW. Effects of frontal-executive dysfunction on self-perceived hearing handicap in the elderly with mild cognitive impairment. PLoS One 2019; 14:e0210014. [PMID: 30840623 PMCID: PMC6402624 DOI: 10.1371/journal.pone.0210014] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2018] [Accepted: 12/14/2018] [Indexed: 02/07/2023] Open
Abstract
It is increasingly agreed upon that cognitive and audiological factors are associated with self-perceived hearing handicap in old adults. This study aimed to compare self-perceived hearing handicap among mild cognitive impairment (MCI) subgroups and a cognitively normal elderly (CNE) group and determine which factors (i.e., demographic, audiometric, or neuropsychological factors) are correlated with self-perceived hearing handicap in each group. A total of 46 MCI patients and 39 hearing threshold-matched CNE subjects participated in this study, and their age ranged from 55 to 80 years. The MCI patients were reclassified into two groups: 16 with frontal-executive dysfunction (FED) and 30 without FED. All subjects underwent audiometric, neuropsychological, and self-perceived hearing handicap assessments. The Korean version of the Hearing Handicap Inventory for the Elderly (K-HHIE) was administered to obtain the hearing handicap scores for each subject. After controlling for age, years of education, and depression levels, we found no significant differences in the K-HHIE scores between the MCI and the CNE groups. However, after we classified the MCI patients into the MCI with FED and MCI without FED groups, the MCI with FED group scored significantly higher than did both the MCI without FED and the CNE groups. In addition, after controlling for depression levels, significant partial correlations of hearing handicap scores with frontal-executive function scores and speech-in-noise perception performance were found in the MCI groups. In the CNE group, the hearing handicap scores were related to peripheral hearing sensitivity and years of education. In summary, MCI patients with FED are more likely to experience everyday hearing handicap than those without FED and cognitively normal old adults. Although educational level and peripheral hearing function are related to self-perceived hearing handicap in cognitively normal old adults, speech-in-noise perception and frontal-executive function are mainly associated with hearing handicap in patients with MCI.
Collapse
Affiliation(s)
- Soo Jung Lee
- Graduate Program in Speech and Language Pathology, Yonsei University, Seoul, Korea
| | - HyangHee Kim
- Graduate Program in Speech and Language Pathology, Yonsei University, Seoul, Korea
- Department and Research Institute of Rehabilitation Medicine, Yonsei University College of Medicine, Seoul, Korea
| | - Lee-Suk Kim
- Department of Otolaryngology, Head and Neck Surgery, Dong-A University College of Medicine, Busan, Korea
| | | | - Kyung Won Park
- Department of Neurology, Dong-A University College of Medicine, Busan, Korea
- Institute of Convergence Bio-Health, Dong-A University, Busan, Korea
- * E-mail:
| |
Collapse
|
35
|
van Schoonhoven J, Rhebergen KS, Dreschler WA. The Extended Speech Transmission Index: Predicting speech intelligibility in fluctuating noise and reverberant rooms. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2019; 145:1178. [PMID: 31067918 DOI: 10.1121/1.5092204] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/25/2018] [Accepted: 02/07/2019] [Indexed: 06/09/2023]
Abstract
The Speech Transmission Index (STI) is used to predict speech intelligibility in noise and reverberant environments. However, measurements and predictions in fluctuating noises lead to inaccuracies. In the current paper, the Extended Speech Transmission Index (ESTI) is presented in order to deal with these shortcomings. Speech intelligibility in normally hearing subjects was measured using stationary and fluctuating maskers. These results served to optimize model parameters. Data from the literature were then used to verify the ESTI-model. Model outcomes were accurate for stationary maskers, maskers with artificial fluctuations, and maskers with real life non-speech modulations. Maskers with speech-like characteristics introduced systematic errors in the model outcomes, probably due to a combination of modulation masking, context effects, and informational masking.
Collapse
Affiliation(s)
- Jelmer van Schoonhoven
- Department of Clinical and Experimental Audiology, Academic Medical Center, Meibergdreef 9, 1105 AZ Amsterdam, The Netherlands
| | - Koenraad S Rhebergen
- Department of Otorhinolaryngology and Head and Neck Surgery, Rudolf Magnus Institute of Neuroscience, University Medical Center, Utrecht, The Netherlands
| | - Wouter A Dreschler
- Department of Clinical and Experimental Audiology, Academic Medical Center, Meibergdreef 9, 1105 AZ Amsterdam, The Netherlands
| |
Collapse
|
36
|
Zekveld AA, Koelewijn T, Kramer SE. The Pupil Dilation Response to Auditory Stimuli: Current State of Knowledge. Trends Hear 2019; 22:2331216518777174. [PMID: 30249172 PMCID: PMC6156203 DOI: 10.1177/2331216518777174] [Citation(s) in RCA: 113] [Impact Index Per Article: 22.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023] Open
Abstract
The measurement of cognitive resource allocation during listening, or listening effort, provides valuable insight in the factors influencing auditory processing. In recent years, many studies inside and outside the field of hearing science have measured the pupil response evoked by auditory stimuli. The aim of the current review was to provide an exhaustive overview of these studies. The 146 studies included in this review originated from multiple domains, including hearing science and linguistics, but the review also covers research into motivation, memory, and emotion. The present review provides a unique overview of these studies and is organized according to the components of the Framework for Understanding Effortful Listening. A summary table presents the sample characteristics, an outline of the study design, stimuli, the pupil parameters analyzed, and the main findings of each study. The results indicate that the pupil response is sensitive to various task manipulations as well as interindividual differences. Many of the findings have been replicated. Frequent interactions between the independent factors affecting the pupil response have been reported, which indicates complex processes underlying cognitive resource allocation. This complexity should be taken into account in future studies that should focus more on interindividual differences, also including older participants. This review facilitates the careful design of new studies by indicating the factors that should be controlled for. In conclusion, measuring the pupil dilation response to auditory stimuli has been demonstrated to be sensitive method applicable to numerous research questions. The sensitivity of the measure calls for carefully designed stimuli.
Collapse
Affiliation(s)
- Adriana A Zekveld
- 1 Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery, Amsterdam Public Health Research Institute, VU University Medical Center, the Netherlands.,2 Linnaeus Centre HEAD, The Swedish Institute for Disability Research, Sweden.,3 Department of Behavioural Sciences and Learning, Linköping University, Sweden
| | - Thomas Koelewijn
- 1 Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery, Amsterdam Public Health Research Institute, VU University Medical Center, the Netherlands
| | - Sophia E Kramer
- 1 Section Ear & Hearing, Department of Otolaryngology-Head and Neck Surgery, Amsterdam Public Health Research Institute, VU University Medical Center, the Netherlands
| |
Collapse
|
37
|
Rönnberg J, Holmer E, Rudner M. Cognitive hearing science and ease of language understanding. Int J Audiol 2019; 58:247-261. [DOI: 10.1080/14992027.2018.1551631] [Citation(s) in RCA: 52] [Impact Index Per Article: 10.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/30/2023]
Affiliation(s)
- Jerker Rönnberg
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, The Swedish Institute for Disability Research, Linköping University, Linköping, Sweden
| | - Emil Holmer
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, The Swedish Institute for Disability Research, Linköping University, Linköping, Sweden
| | - Mary Rudner
- Department of Behavioural Sciences and Learning, Linnaeus Centre HEAD, The Swedish Institute for Disability Research, Linköping University, Linköping, Sweden
| |
Collapse
|
38
|
Toward a more comprehensive understanding of the impact of masker type and signal-to-noise ratio on the pupillary response while performing a speech-in-noise test. Hear Res 2018; 369:67-78. [DOI: 10.1016/j.heares.2018.05.006] [Citation(s) in RCA: 40] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/03/2017] [Revised: 04/07/2018] [Accepted: 05/09/2018] [Indexed: 11/19/2022]
|
39
|
McCloy DR, Larson E, Lee AKC. Auditory attention switching with listening difficulty: Behavioral and pupillometric measures. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2018; 144:2764. [PMID: 30522295 PMCID: PMC6232045 DOI: 10.1121/1.5078618] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/13/2023]
Abstract
Pupillometry has emerged as a useful tool for studying listening effort. Past work involving listeners with normal audiological thresholds has shown that switching attention between competing talker streams evokes pupil dilation indicative of listening effort [McCloy, Lau, Larson, Pratt, and Lee (2017). J. Acoust. Soc. Am. 141(4), 2440-2451]. The current experiment examines behavioral and pupillometric data from a two-stream target detection task requiring attention-switching between auditory streams, in two participant groups: audiometrically normal listeners who self-report difficulty localizing sound sources and/or understanding speech in reverberant or acoustically crowded environments, and their age-matched controls who do not report such problems. Three experimental conditions varied the number and type of stream segregation cues available. Participants who reported listening difficulty showed both behavioral and pupillometric signs of increased effort compared to controls, especially in trials where listeners had to switch attention between streams, or trials where only a single stream segregation cue was available.
Collapse
Affiliation(s)
- Daniel R McCloy
- Institute for Learning and Brain Sciences, 1715 NE Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Eric Larson
- Institute for Learning and Brain Sciences, 1715 NE Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| | - Adrian K C Lee
- Institute for Learning and Brain Sciences, 1715 NE Columbia Road, Box 357988, Seattle, Washington 98195-7988, USA
| |
Collapse
|
40
|
Lee SJ, Park KW, Kim LS, Kim H. Association between Frontal-Executive Dysfunction and Speech-in-Noise Perception Deficits in Mild Cognitive Impairment. J Clin Neurol 2018; 14:513-522. [PMID: 30198228 PMCID: PMC6172502 DOI: 10.3988/jcn.2018.14.4.513] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2018] [Revised: 05/21/2018] [Accepted: 05/23/2018] [Indexed: 12/22/2022] Open
Abstract
Background and Purpose Speech-in-noise perception deficits have been demonstrated in patients with mild cognitive impairment (MCI). However, it remains unclear whether the impairment of speech perception varies between MCI subtypes. The purpose of this study was twofold: 1) to compare speech perception performance among MCI subgroups, and 2) to identify the cognitive domains specifically related to speech-in-noise perception. Methods We studied 46 patients with MCI and 39 hearing-threshold-matched cognitively normal elderly (CNE) subjects. Two different patient classifications were used: 1) patients with amnestic mild cognitive impairment (aMCI) (n=21) or nonamnestic mild cognitive impairment (naMCI) (n=25), and 2) patients with frontal-executive dysfunction (FED) (n=16) or without FED (n=30). All of the subjects underwent audiometric, neuropsychological, and speech perception assessments. Speech-in-noise perception was measured using sentence recognition tests in the presence of two types of background noise at four levels. Results First, as the level of background noise increased, the MCI with FED group scored lower than both the MCI without FED and CNE groups under both types of noise. Second, both the naMCI and aMCI groups scored lower than the CNE group, but there were no differences between the naMCI and aMCI groups in sentence recognition under any noise conditions. Third, significant correlations were found between sentence recognition and executive function scores both in the MCI groups and in the CNE group. Conclusions Our findings suggest that frontal-executive function is strongly related to speech-in-noise perception and that MCI patients with FED have greater deficits in speech-in-noise perception compared to other subgroups of MCI.
Collapse
Affiliation(s)
- Soo Jung Lee
- Department of Audiology and Speech-Language Pathology, Daegu Catholic University, Gyeongsan, Korea
| | - Kyung Won Park
- Department of Neurology, Dong-A University College of Medicine, Busan, Korea
| | - Lee Suk Kim
- Department of Otolaryngology, Head and Neck Surgery, Dong-A University College of Medicine, Busan, Korea
| | - HyangHee Kim
- Graduate Program in Speech-Language Pathology and Department and Research Institute of Rehabilitation Medicine, Yonsei University College of Medicine, Seoul, Korea.
| |
Collapse
|
41
|
The effect of reward on listening effort as reflected by the pupil dilation response. Hear Res 2018; 367:106-112. [DOI: 10.1016/j.heares.2018.07.011] [Citation(s) in RCA: 35] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/02/2018] [Revised: 07/19/2018] [Accepted: 07/25/2018] [Indexed: 11/22/2022]
|
42
|
Psychophysiological measurement of affective responses during speech perception. Hear Res 2018; 369:103-119. [PMID: 30135023 DOI: 10.1016/j.heares.2018.07.007] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/31/2017] [Revised: 07/02/2018] [Accepted: 07/13/2018] [Indexed: 02/07/2023]
Abstract
When people make decisions about listening, such as whether to continue attending to a particular conversation or whether to wear their hearing aids to a particular restaurant, they do so on the basis of more than just their estimated performance. Recent research has highlighted the vital role of more subjective qualities such as effort, motivation, and fatigue. Here, we argue that the importance of these factors is largely mediated by a listener's emotional response to the listening challenge, and suggest that emotional responses to communication challenges may provide a crucial link between day-to-day communication stress and long-term health. We start by introducing some basic concepts from the study of emotion and affect. We then develop a conceptual framework to guide future research on this topic through examination of a variety of autonomic and peripheral physiological responses that have been employed to investigate both cognitive and affective phenomena related to challenging communication. We conclude by suggesting the need for further investigation of the links between communication difficulties, emotional response, and long-term health, and make some recommendations intended to guide future research on affective psychophysiology in speech communication.
Collapse
|
43
|
Francis AL, Tigchelaar LJ, Zhang R, Zekveld AA. Effects of Second Language Proficiency and Linguistic Uncertainty on Recognition of Speech in Native and Nonnative Competing Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:1815-1830. [PMID: 29971338 DOI: 10.1044/2018_jslhr-h-17-0254] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/03/2017] [Accepted: 03/26/2018] [Indexed: 06/08/2023]
Abstract
PURPOSE The purpose of this study was to investigate the effects of 2nd language proficiency and linguistic uncertainty on performance and listening effort in mixed language contexts. METHOD Thirteen native speakers of Dutch with varying degrees of fluency in English listened to and repeated sentences produced in both Dutch and English and presented in the presence of single-talker competing speech in both Dutch and English. Target and masker language combinations were presented in both blocked and mixed (unpredictable) conditions. In the blocked condition, in each block of trials the target-masker language combination remained constant, and the listeners were informed of both prior to beginning the block. In the mixed condition, target and masker language varied randomly from trial to trial. All listeners participated in all conditions. Performance was assessed in terms of speech reception thresholds, whereas listening effort was quantified in terms of pupil dilation. RESULTS Performance (speech reception thresholds) and listening effort (pupil dilation) were both affected by 2nd language proficiency (English test score) and target and masker language: Performance was better in blocked as compared to mixed conditions, with Dutch as compared to English targets, and with English as compared to Dutch maskers. English proficiency was correlated with listening performance. Listeners also exhibited greater peak pupil dilation in mixed as compared to blocked conditions for trials with Dutch maskers, whereas pupil dilation during preparation for speaking was higher for English targets as compared to Dutch ones in almost all conditions. CONCLUSIONS Both listener's proficiency in a 2nd language and uncertainty about the target language on a given trial play a significant role in how bilingual listeners attend to speech in the presence of competing speech in different languages, but precise effects also depend on which language is serving as target and which as masker.
Collapse
Affiliation(s)
- Alexander L Francis
- Department of Speech, Language & Hearing Sciences, Purdue University, West Lafayette, IN
| | | | - Rongrong Zhang
- Department of Statistics, Purdue University, West Lafayette, IN
| | - Adriana A Zekveld
- VU University Medical Center, Amsterdam, the Netherlands
- Linnaeus Centre, Linköping University, Sweden
| |
Collapse
|
44
|
Strand JF, Brown VA, Merchant MB, Brown HE, Smith J. Measuring Listening Effort: Convergent Validity, Sensitivity, and Links With Cognitive and Personality Measures. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:1463-1486. [PMID: 29800081 DOI: 10.1044/2018_jslhr-h-17-0257] [Citation(s) in RCA: 73] [Impact Index Per Article: 12.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/07/2017] [Accepted: 02/06/2018] [Indexed: 06/08/2023]
Abstract
PURPOSE Listening effort (LE) describes the attentional or cognitive requirements for successful listening. Despite substantial theoretical and clinical interest in LE, inconsistent operationalization makes it difficult to make generalizations across studies. The aims of this large-scale validation study were to evaluate the convergent validity and sensitivity of commonly used measures of LE and assess how scores on those tasks relate to cognitive and personality variables. METHOD Young adults with normal hearing (N = 111) completed 7 tasks designed to measure LE, 5 tests of cognitive ability, and 2 personality measures. RESULTS Scores on some behavioral LE tasks were moderately intercorrelated but were generally not correlated with subjective and physiological measures of LE, suggesting that these tasks may not be tapping into the same underlying construct. LE measures differed in their sensitivity to changes in signal-to-noise ratio and the extent to which they correlated with cognitive and personality variables. CONCLUSIONS Given that LE measures do not show consistent, strong intercorrelations and differ in their relationships with cognitive and personality predictors, these findings suggest caution in generalizing across studies that use different measures of LE. The results also indicate that people with greater cognitive ability appear to use their resources more efficiently, thereby diminishing the detrimental effects associated with increased background noise during language processing.
Collapse
Affiliation(s)
- Julia F Strand
- Department of Psychology, Carleton College, Northfield, MN
| | - Violet A Brown
- Department of Psychology, Carleton College, Northfield, MN
| | | | - Hunter E Brown
- Department of Psychology, Carleton College, Northfield, MN
| | - Julia Smith
- Department of Psychology, Carleton College, Northfield, MN
| |
Collapse
|
45
|
Nuesse T, Steenken R, Neher T, Holube I. Exploring the Link Between Cognitive Abilities and Speech Recognition in the Elderly Under Different Listening Conditions. Front Psychol 2018; 9:678. [PMID: 29867654 PMCID: PMC5968383 DOI: 10.3389/fpsyg.2018.00678] [Citation(s) in RCA: 27] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2017] [Accepted: 04/19/2018] [Indexed: 12/20/2022] Open
Abstract
Elderly listeners are known to differ considerably in their ability to understand speech in noise. Several studies have addressed the underlying factors that contribute to these differences. These factors include audibility, and age-related changes in supra-threshold auditory processing abilities, and it has been suggested that differences in cognitive abilities may also be important. The objective of this study was to investigate associations between performance in cognitive tasks and speech recognition under different listening conditions in older adults with either age appropriate hearing or hearing-impairment. To that end, speech recognition threshold (SRT) measurements were performed under several masking conditions that varied along the perceptual dimensions of dip listening, spatial separation, and informational masking. In addition, a neuropsychological test battery was administered, which included measures of verbal working and short-term memory, executive functioning, selective and divided attention, and lexical and semantic abilities. Age-matched groups of older adults with either age-appropriate hearing (ENH, n = 20) or aided hearing impairment (EHI, n = 21) participated. In repeated linear regression analyses, composite scores of cognitive test outcomes (evaluated using PCA) were included to predict SRTs. These associations were different for the two groups. When hearing thresholds were controlled for, composed cognitive factors were significantly associated with the SRTs for the ENH listeners. Whereas better lexical and semantic abilities were associated with lower (better) SRTs in this group, there was a negative association between attentional abilities and speech recognition in the presence of spatially separated speech-like maskers. For the EHI group, the pure-tone thresholds (averaged across 0.5, 1, 2, and 4 kHz) were significantly associated with the SRTs, despite the fact that all signals were amplified and therefore in principle audible.
Collapse
Affiliation(s)
- Theresa Nuesse
- Institute of Hearing Technology and Audiology, Jade University of Applied Sciences, Oldenburg, Germany.,Cluster of Excellence "Hearing4All", Oldenburg, Germany
| | - Rike Steenken
- Institute of Hearing Technology and Audiology, Jade University of Applied Sciences, Oldenburg, Germany.,Cluster of Excellence "Hearing4All", Oldenburg, Germany
| | - Tobias Neher
- Cluster of Excellence "Hearing4All", Oldenburg, Germany.,Medizinische Physik, Oldenburg University, Oldenburg, Germany.,Faculty of Health Sciences, Institute of Clinical Research, University of Southern Denmark, Odense, Denmark
| | - Inga Holube
- Institute of Hearing Technology and Audiology, Jade University of Applied Sciences, Oldenburg, Germany.,Cluster of Excellence "Hearing4All", Oldenburg, Germany
| |
Collapse
|
46
|
|
47
|
van Knijff EC, Coene M, Govaerts PJ. Speech understanding in noise in elderly adults: the effect of inhibitory control and syntactic complexity. INTERNATIONAL JOURNAL OF LANGUAGE & COMMUNICATION DISORDERS 2018; 53:628-642. [PMID: 29446191 DOI: 10.1111/1460-6984.12376] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/01/2017] [Revised: 01/12/2018] [Accepted: 01/16/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND Previous research has suggested that speech perception in elderly adults is influenced not only by age-related hearing loss or presbycusis but also by declines in cognitive abilities, by background noise and by the syntactic complexity of the message. AIMS To gain further insight into the influence of these cognitive as well as acoustic and linguistic factors on speech perception in elderly adults by investigating inhibitory control as a listener characteristic and background noise type and syntactic complexity as input characteristics. METHODS & PROCEDURES Phoneme identification was measured in different noise conditions and in different linguistic contexts (single words, sentences with varying syntactic complexity). Additionally, inhibitory control was measured using a visual stimulus-response matching task. Fifty-one adults participated in this study, including elderly adults with age-related hearing loss (n = 9) and with normal hearing (n = 17), and a control group of normal hearing younger adults (n = 25). OUTCOMES & RESULTS The analysis revealed that elderly adults with normal hearing and with hearing loss were less likely to identify successfully phonemes in single words than younger normal hearing controls. In the context of sentences, only elderly adults with hearing loss had a lower odds of correct phoneme perception than the control group. Additionally, in elderly adults with hearing loss, phoneme-in-sentence perception was linked to age-related declines in inhibitory control. In all participants, phoneme identification in sentences was influenced by both noise type and syntactic complexity. CONCLUSIONS & IMPLICATIONS Inhibitory control and syntactic complexity might play a significant role in speech perception, especially in elderly listeners. These factors might also influence the results of clinical assessments of speech perception. Testing procedures thus need to be selected and their results interpreted carefully with these influences in mind.
Collapse
Affiliation(s)
- Eline C van Knijff
- Language and Hearing Center Amsterdam, Vrije Universiteit Amsterdam, the Netherlands
| | - Martine Coene
- Language and Hearing Center Amsterdam, Vrije Universiteit Amsterdam, the Netherlands
- The Eargroup, Antwerp, Belgium
| | - Paul J Govaerts
- Language and Hearing Center Amsterdam, Vrije Universiteit Amsterdam, the Netherlands
- The Eargroup, Antwerp, Belgium
| |
Collapse
|
48
|
Miles K, McMahon C, Boisvert I, Ibrahim R, de Lissa P, Graham P, Lyxell B. Objective Assessment of Listening Effort: Coregistration of Pupillometry and EEG. Trends Hear 2018; 21:2331216517706396. [PMID: 28752807 PMCID: PMC5536372 DOI: 10.1177/2331216517706396] [Citation(s) in RCA: 36] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/01/2023] Open
Abstract
Listening to speech in noise is effortful, particularly for people with hearing impairment. While it is known that effort is related to a complex interplay between bottom-up and top-down processes, the cognitive and neurophysiological mechanisms contributing to effortful listening remain unknown. Therefore, a reliable physiological measure to assess effort remains elusive. This study aimed to determine whether pupil dilation and alpha power change, two physiological measures suggested to index listening effort, assess similar processes. Listening effort was manipulated by parametrically varying spectral resolution (16- and 6-channel noise vocoding) and speech reception thresholds (SRT; 50% and 80%) while 19 young, normal-hearing adults performed a speech recognition task in noise. Results of off-line sentence scoring showed discrepancies between the target SRTs and the true performance obtained during the speech recognition task. For example, in the SRT80% condition, participants scored an average of 64.7%. Participants’ true performance levels were therefore used for subsequent statistical modelling. Results showed that both measures appeared to be sensitive to changes in spectral resolution (channel vocoding), while pupil dilation only was also significantly related to their true performance levels (%) and task accuracy (i.e., whether the response was correctly or partially recalled). The two measures were not correlated, suggesting they each may reflect different cognitive processes involved in listening effort. This combination of findings contributes to a growing body of research aiming to develop an objective measure of listening effort.
Collapse
Affiliation(s)
- Kelly Miles
- 1 Department of Linguistics, Australian Hearing Hub, Macquarie University, Sydney, Australia.,2 The HEARing Cooperative Research Centre, Melbourne, Australia.,3 Linnaeus Centre for HEaring And Deafness (HEAD), Swedish Institute for Disability Research, Linköping University, Sweden
| | - Catherine McMahon
- 1 Department of Linguistics, Australian Hearing Hub, Macquarie University, Sydney, Australia.,2 The HEARing Cooperative Research Centre, Melbourne, Australia
| | - Isabelle Boisvert
- 1 Department of Linguistics, Australian Hearing Hub, Macquarie University, Sydney, Australia.,2 The HEARing Cooperative Research Centre, Melbourne, Australia
| | - Ronny Ibrahim
- 1 Department of Linguistics, Australian Hearing Hub, Macquarie University, Sydney, Australia.,2 The HEARing Cooperative Research Centre, Melbourne, Australia
| | - Peter de Lissa
- 2 The HEARing Cooperative Research Centre, Melbourne, Australia.,4 Department of Psychology, Australian Hearing Hub, Macquarie University, Sydney, Australia
| | - Petra Graham
- 5 Department of Statistics, Australian Hearing Hub, Macquarie University, Sydney, Australia
| | - Björn Lyxell
- 3 Linnaeus Centre for HEaring And Deafness (HEAD), Swedish Institute for Disability Research, Linköping University, Sweden
| |
Collapse
|
49
|
Zekveld AA, Pronk M, Danielsson H, Rönnberg J. Reading Behind the Lines: The Factors Affecting the Text Reception Threshold in Hearing Aid Users. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2018; 61:762-775. [PMID: 29450534 DOI: 10.1044/2017_jslhr-h-17-0196] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/23/2017] [Accepted: 10/12/2017] [Indexed: 06/08/2023]
Abstract
PURPOSE The visual Text Reception Threshold (TRT) test (Zekveld et al., 2007) has been designed to assess modality-general factors relevant for speech perception in noise. In the last decade, the test has been adopted in audiology labs worldwide. The 1st aim of this study was to examine which factors best predict interindividual differences in the TRT. Second, we aimed to assess the relationships between the TRT and the speech reception thresholds (SRTs) estimated in various conditions. METHOD First, we reviewed studies reporting relationships between the TRT and the auditory and/or cognitive factors and formulated specific hypotheses regarding the TRT predictors. These hypotheses were tested using a prediction model applied to a rich data set of 180 hearing aid users. In separate association models, we tested the relationships between the TRT and the various SRTs and subjective hearing difficulties, while taking into account potential confounding variables. RESULTS The results of the prediction model indicate that the TRT is predicted by the ability to fill in missing words in incomplete sentences, by lexical access speed, and by working memory capacity. Furthermore, in line with previous studies, a moderate association between higher age, poorer pure-tone hearing acuity, and poorer TRTs was observed. Better TRTs were associated with better SRTs for the correct perception of 50% of Hagerman matrix sentences in a 4-talker babble, as well as with better subjective ratings of speech perception. Age and pure-tone hearing thresholds significantly confounded these associations. The associations of the TRT with SRTs estimated in other conditions and with subjective qualities of hearing were not statistically significant when adjusting for age and pure-tone average. CONCLUSIONS We conclude that the abilities tapped into by the TRT test include processes relevant for speeded lexical decision making when completing partly masked sentences and that these processes require working memory capacity. Furthermore, the TRT is associated with the SRT of hearing aid users as estimated in a challenging condition that includes informational masking and with experienced difficulties with speech perception in daily-life conditions. The current results underline the value of using the TRT test in studies involving speech perception and aid in the interpretation of findings acquired using the test.
Collapse
Affiliation(s)
- Adriana A Zekveld
- Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Linnaeus Centre HEAD, Swedish Institute for Disability Research, Linköping, Sweden
- Section Ear & Hearing, Department of Otolaryngology/Head & Neck Surgery, Amsterdam Public Health Research Institute, VU University Medical Center, the Netherlands
| | - Marieke Pronk
- Section Ear & Hearing, Department of Otolaryngology/Head & Neck Surgery, Amsterdam Public Health Research Institute, VU University Medical Center, the Netherlands
| | - Henrik Danielsson
- Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Linnaeus Centre HEAD, Swedish Institute for Disability Research, Linköping, Sweden
| | - Jerker Rönnberg
- Department of Behavioural Sciences and Learning, Linköping University, Sweden
- Linnaeus Centre HEAD, Swedish Institute for Disability Research, Linköping, Sweden
| |
Collapse
|
50
|
Borghini G, Hazan V. Listening Effort During Sentence Processing Is Increased for Non-native Listeners: A Pupillometry Study. Front Neurosci 2018; 12:152. [PMID: 29593489 PMCID: PMC5859302 DOI: 10.3389/fnins.2018.00152] [Citation(s) in RCA: 37] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2017] [Accepted: 02/26/2018] [Indexed: 11/24/2022] Open
Abstract
Current evidence demonstrates that even though some non-native listeners can achieve native-like performance for speech perception tasks in quiet, the presence of a background noise is much more detrimental to speech intelligibility for non-native compared to native listeners. Even when performance is equated across groups, it is likely that greater listening effort is required for non-native listeners. Importantly, the added listening effort might result in increased fatigue and a reduced ability to successfully perform multiple tasks simultaneously. Task-evoked pupil responses have been demonstrated to be a reliable measure of cognitive effort and can be useful in clarifying those aspects. In this study we compared the pupil response for 23 native English speakers and 27 Italian speakers of English as a second language. Speech intelligibility was tested for sentences presented in quiet and in background noise at two performance levels that were matched across groups. Signal-to-noise levels corresponding to these sentence intelligibility levels were pre-determined using an adaptive intelligibility task. Pupil response was significantly greater in non-native compared to native participants across both intelligibility levels. Therefore, for a given intelligibility level, a greater listening effort is required when listening in a second language in order to understand speech in noise. Results also confirmed that pupil response is sensitive to speech intelligibility during language comprehension, in line with previous research. However, contrary to our predictions, pupil response was not differentially modulated by intelligibility levels for native and non-native listeners. The present study corroborates that pupillometry can be deemed as a valid measure to be used in speech perception investigation, because it is sensitive to differences both across participants, such as listener type, and across conditions, such as variations in the level of speech intelligibility. Importantly, pupillometry offers us the possibility to uncover differences in listening effort even when those do not emerge in the performance level of individuals.
Collapse
Affiliation(s)
- Giulia Borghini
- Department of Speech Hearing and Phonetic Sciences, Faculty of Brain Sciences, University College London, London, United Kingdom
| | - Valerie Hazan
- Department of Speech Hearing and Phonetic Sciences, Faculty of Brain Sciences, University College London, London, United Kingdom
| |
Collapse
|