1
|
Harel-Arbeli T, Shaposhnik H, Palgi Y, Ben-David BM. Taking the Extra Listening Mile: Processing Spoken Semantic Context Is More Effortful for Older Than Young Adults. Ear Hear 2025; 46:315-324. [PMID: 39219019 DOI: 10.1097/aud.0000000000001582] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
OBJECTIVES Older adults use semantic context to generate predictions in speech processing, compensating for aging-related sensory and cognitive changes. This study aimed to gauge aging-related changes in effort exertion related to context use. DESIGN The study revisited data from Harel-Arbeli et al. (2023) that used a "visual-world" eye-tracking paradigm. Data on efficiency of context use (response latency and the probability to gaze at the target before hearing it) and effort exertion (pupil dilation) were extracted from a subset of 14 young adults (21 to 27 years old) and 13 older adults (65 to 79 years old). RESULTS Both age groups showed a similar pattern of context benefits for response latency and target word predictions, however only the older adults group showed overall increased pupil dilation when listening to context sentences. CONCLUSIONS Older adults' efficient use of spoken semantic context appears to come at a cost of increased effort exertion.
Collapse
Affiliation(s)
- Tami Harel-Arbeli
- Department of Gerontology, Haifa University, Haifa, Israel
- Communication, Aging and Neuropsychology Lab, Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- Department of Communication Disorders, Achva Academic College, Arugot, Israel
| | - Hagit Shaposhnik
- Department of Software and Information Systems Engineering, Ben-Gurion University of the Negev, Beer-Sheva, Israel
| | - Yuval Palgi
- Department of Gerontology, Haifa University, Haifa, Israel
| | - Boaz M Ben-David
- Communication, Aging and Neuropsychology Lab, Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- KITE, Toronto Rehabilitation Institute, University Health Networks, Toronto, Ontario, Canada
- Department of Speech-Language Pathology, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
2
|
Mansour A, Ben-David BM, Sasson A, Farraj J, Mansour A, Roth Y, Icht M. Association between oral feeding versus enteral feeding and cerumen impaction in older hospitalized adults: A retrospective cohort study. JPEN J Parenter Enteral Nutr 2025. [PMID: 39829405 DOI: 10.1002/jpen.2724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Revised: 11/27/2024] [Accepted: 12/29/2024] [Indexed: 01/22/2025]
Abstract
BACKGROUND Chewing involves jaw movements that propel cerumen along the ear canal. This mechanism may be reduced in dysphagia, especially for older individuals who are enterally fed. Those patients may be at a higher risk for cerumen impaction and may require longer hospital stays. Examining the relationship between diet type, cerumen impaction, and hospital stay duration was the focus of the present study. METHODS We performed a retrospective cohort study (not registered) among 114 hospitalized older adults. Data were collected on diet type: (1) oral feeding (individuals fed a solid diet or a pureed diet) or (2) enteral feeding (individuals fed via a feeding tube). The results of an otoscopy that quantified cerumen were recorded, as well as hospital stay duration. RESULTS In a mediation analysis, a hospital stay of >1 month was associated with an increased risk of enteral feeding, which in turn, increased the risk of cerumen impaction. Analysis indicated that the link between longer hospitalization and a more severe level of cerumen impaction was fully mediated by diet type (enteral feeding). CONCLUSIONS Enteral feeding seems to be a risk factor for cerumen impaction, rather than merely hospitalization length, in our sample of geriatric patients. These results highlight the importance of continuous monitoring by ear, nose, and throat specialists, as well as regular auditory assessments for patients who are enterally fed for early detection and treatment of cerumen impaction. Particular attention should be paid to cases of prolonged hospitalization, which is associated with the severity of dysphagia.
Collapse
Affiliation(s)
- Asil Mansour
- Communication Disorders Unit, Dorot Geriatric Medical Rehabilitative Center, Netanya, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- Department of Speech-Language Pathology, University of Toronto, Toronto, Ontario, Canada
- KITE, Toronto Rehabilitation Institute, University Health Networks (UHN), Toronto, Ontario, Canada
| | - Ady Sasson
- Dorot Geriatric Medical Rehabilitative Center, Netanya, Israel
| | - Jalal Farraj
- Dorot Geriatric Medical Rehabilitative Center, Netanya, Israel
| | - Anwar Mansour
- Department of Otolaryngology, Meir Medical Center, Kfar Saba, Israel
- Sackler School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | | | - Michal Icht
- Department of Communication Disorders, Ariel University, Ariel, Israel
| |
Collapse
|
3
|
Abdel-Latif KHA, Koelewijn T, Başkent D, Meister H. Assessment of Speech Processing and Listening Effort Associated With Speech-on-Speech Masking Using the Visual World Paradigm and Pupillometry. Trends Hear 2025; 29:23312165241306091. [PMID: 39800920 PMCID: PMC11726529 DOI: 10.1177/23312165241306091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2024] [Revised: 10/16/2024] [Accepted: 11/23/2024] [Indexed: 01/16/2025] Open
Abstract
Speech-on-speech masking is a common and challenging situation in everyday verbal communication. The ability to segregate competing auditory streams is a necessary requirement for focusing attention on the target speech. The Visual World Paradigm (VWP) provides insight into speech processing by capturing gaze fixations on visually presented icons that reflect the speech signal. This study aimed to propose a new VWP to examine the time course of speech segregation when competing sentences are presented and to collect pupil size data as a measure of listening effort. Twelve young normal-hearing participants were presented with competing matrix sentences (structure "name-verb-numeral-adjective-object") diotically via headphones at four target-to-masker ratios (TMRs), corresponding to intermediate to near perfect speech recognition. The VWP visually presented the number and object words from both the target and masker sentences. Participants were instructed to gaze at the corresponding words of the target sentence without providing verbal responses. The gaze fixations consistently reflected the different TMRs for both number and object words. The slopes of the fixation curves were steeper, and the proportion of target fixations increased with higher TMRs, suggesting more efficient segregation under more favorable conditions. Temporal analysis of pupil data using Bayesian paired sample t-tests showed a corresponding reduction in pupil dilation with increasing TMR, indicating reduced listening effort. The results support the conclusion that the proposed VWP and the captured eye movements and pupil dilation are suitable for objective assessment of sentence-based speech-on-speech segregation and the corresponding listening effort.
Collapse
Affiliation(s)
- Khaled H. A. Abdel-Latif
- Faculty of Medicine and University Hospital Cologne, Department of Otorhinolaryngology, Head and Neck Surgery, University of Cologne, Cologne, Germany
- Jean Uhrmacher Institute for Clinical ENT-Research, University of Cologne, Cologne, Germany
| | - Thomas Koelewijn
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands
| | - Deniz Başkent
- Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands
- Research School of Behavioural and Cognitive Neurosciences, Graduate School of Medical Sciences, University of Groningen, Groningen, The Netherlands
| | - Hartmut Meister
- Faculty of Medicine and University Hospital Cologne, Department of Otorhinolaryngology, Head and Neck Surgery, University of Cologne, Cologne, Germany
- Jean Uhrmacher Institute for Clinical ENT-Research, University of Cologne, Cologne, Germany
| |
Collapse
|
4
|
Abbas M, Szpiro SFA, Karawani H. Interconnected declines in audition vision and cognition in healthy aging. Sci Rep 2024; 14:30805. [PMID: 39730569 DOI: 10.1038/s41598-024-81154-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Accepted: 11/25/2024] [Indexed: 12/29/2024] Open
Abstract
Age-related sensory declines are unavoidable and closely linked to decreased visual, auditory, and cognitive functions. However, the interrelations of these declines remain poorly understood. Despite extensive studies in each domain, shared age-related characteristics are complex and may not consistently manifest direct relationships at the individual level. We investigated the link between visual and auditory perceptual declines in healthy aging and their relation to cognitive function using six psychophysical and three cognitive tasks. Eighty young and older healthy adults participated, revealing a general age-related decline. Young adults consistently outperformed older adults in all tasks. Critically, the performance in visual tasks significantly correlated with performance in auditory tasks in older adults. This suggests a domain-general decline in perception, where declines in vision are related to declines in audition within individuals. Additionally, perceptual performance in older adults decreased monotonically year by year. Working memory performance significantly correlated with perceptual performance across both age groups and modalities, further supporting the hypothesis of a domain-general decline. These findings highlight the complex and interconnected nature of sensory and cognitive declines in aging, providing a foundation for future translational research focused on enhancing cognitive and perceptual abilities to promote healthy aging and ultimately improve the quality of life for older adults.
Collapse
Affiliation(s)
- Mais Abbas
- Department of Communication Sciences and Disorders, Faculty of Social Welfare and Health Sciences, University of Haifa, Haifa, Israel
| | - Sarit F A Szpiro
- Department of Special Education, University of Haifa, Haifa, Israel
- Edmond J. Safra Brain Research Center, University of Haifa, Haifa, Israel
- The Haifa Brain and Behavior Hub, University of Haifa, Haifa, Israel
| | - Hanin Karawani
- Department of Communication Sciences and Disorders, Faculty of Social Welfare and Health Sciences, University of Haifa, Haifa, Israel.
- The Haifa Brain and Behavior Hub, University of Haifa, Haifa, Israel.
| |
Collapse
|
5
|
Taitelbaum-Swead R, Ben-David BM. The Role of Early Intact Auditory Experience on the Perception of Spoken Emotions, Comparing Prelingual to Postlingual Cochlear Implant Users. Ear Hear 2024; 45:1585-1599. [PMID: 39004788 DOI: 10.1097/aud.0000000000001550] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/16/2024]
Abstract
OBJECTIVES Cochlear implants (CI) are remarkably effective, but have limitations regarding the transformation of the spectro-temporal fine structures of speech. This may impair processing of spoken emotions, which involves the identification and integration of semantic and prosodic cues. Our previous study found spoken-emotions-processing differences between CI users with postlingual deafness (postlingual CI) and normal hearing (NH) matched controls (age range, 19 to 65 years). Postlingual CI users over-relied on semantic information in incongruent trials (prosody and semantics present different emotions), but rated congruent trials (same emotion) similarly to controls. Postlingual CI's intact early auditory experience may explain this pattern of results. The present study examined whether CI users without intact early auditory experience (prelingual CI) would generally perform worse on spoken emotion processing than NH and postlingual CI users, and whether CI use would affect prosodic processing in both CI groups. First, we compared prelingual CI users with their NH controls. Second, we compared the results of the present study to our previous study ( Taitlebaum-Swead et al. 2022 ; postlingual CI). DESIGN Fifteen prelingual CI users and 15 NH controls (age range, 18 to 31 years) listened to spoken sentences composed of different combinations (congruent and incongruent) of three discrete emotions (anger, happiness, sadness) and neutrality (performance baseline), presented in prosodic and semantic channels (Test for Rating of Emotions in Speech paradigm). Listeners were asked to rate (six-point scale) the extent to which each of the predefined emotions was conveyed by the sentence as a whole (integration of prosody and semantics), or to focus only on one channel (rating the target emotion [RTE]) and ignore the other (selective attention). In addition, all participants performed standard tests of speech perception. Performance on the Test for Rating of Emotions in Speech was compared with the previous study (postlingual CI). RESULTS When asked to focus on one channel, semantics or prosody, both CI groups showed a decrease in prosodic RTE (compared with controls), but only the prelingual CI group showed a decrease in semantic RTE. When the task called for channel integration, both groups of CI users used semantic emotional information to a greater extent than their NH controls. Both groups of CI users rated sentences that did not present the target emotion higher than their NH controls, indicating some degree of confusion. However, only the prelingual CI group rated congruent sentences lower than their NH controls, suggesting reduced accumulation of information across channels. For prelingual CI users, individual differences in identification of monosyllabic words were significantly related to semantic identification and semantic-prosodic integration. CONCLUSIONS Taken together with our previous study, we found that the degradation of acoustic information by the CI impairs the processing of prosodic emotions, in both CI user groups. This distortion appears to lead CI users to over-rely on the semantic information when asked to integrate across channels. Early intact auditory exposure among CI users was found to be necessary for the effective identification of semantic emotions, as well as the accumulation of emotional information across the two channels. Results suggest that interventions for spoken-emotion processing should not ignore the onset of hearing loss.
Collapse
Affiliation(s)
- Riki Taitelbaum-Swead
- Department of Communication Disorders, Speech Perception and Listening Effort Lab in the name of Prof. Mordechai Himelfarb, Ariel University, Israel
- Meuhedet Health Services, Tel Aviv, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
- KITE Research Institute, Toronto Rehabilitation Institute-University Health Network, Toronto, Ontario, Canada
| |
Collapse
|
6
|
McMurray B, Smith FX, Huffman M, Rooff K, Muegge JB, Jeppsen C, Kutlu E, Colby S. Underlying dimensions of real-time word recognition in cochlear implant users. Nat Commun 2024; 15:7382. [PMID: 39209837 PMCID: PMC11362525 DOI: 10.1038/s41467-024-51514-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Accepted: 08/08/2024] [Indexed: 09/04/2024] Open
Abstract
Word recognition is a gateway to language, linking sound to meaning. Prior work has characterized its cognitive mechanisms as a form of competition between similar-sounding words. However, it has not identified dimensions along which this competition varies across people. We sought to identify these dimensions in a population of cochlear implant users with heterogenous backgrounds and audiological profiles, and in a lifespan sample of people without hearing loss. Our study characterizes the process of lexical competition using the Visual World Paradigm. A principal component analysis reveals that people's ability to resolve lexical competition varies along three dimensions that mirror prior small-scale studies. These dimensions capture the degree to which lexical access is delayed ("Wait-and-See"), the degree to which competition fully resolves ("Sustained-Activation"), and the overall rate of activation. Each dimension is predicted by a different auditory skills and demographic factors (onset of deafness, age, cochlear implant experience). Moreover, each dimension predicts outcomes (speech perception in quiet and noise, subjective listening success) over and above auditory fidelity. Higher degrees of Wait-and-See and Sustained-Activation predict poorer outcomes. These results suggest the mechanisms of word recognition vary along a few underlying dimensions which help explain variable performance among listeners encountering auditory challenge.
Collapse
Affiliation(s)
- Bob McMurray
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA.
- Dept. of Communication Sciences & Disorders, University of Iowa, Iowa City, IA, USA.
- Dept. of Otolaryngology-Head and Neck Surgery, University of Iowa, Iowa City, IA, USA.
- Dept. of Linguistics, University of Iowa, Iowa City, IA, USA.
| | - Francis X Smith
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA
- Dept. of Communication Sciences & Disorders, University of Iowa, Iowa City, IA, USA
| | - Marissa Huffman
- Dept. of Otolaryngology-Head and Neck Surgery, University of Iowa, Iowa City, IA, USA
| | - Kristin Rooff
- Dept. of Otolaryngology-Head and Neck Surgery, University of Iowa, Iowa City, IA, USA
| | - John B Muegge
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Charlotte Jeppsen
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA
| | - Ethan Kutlu
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA
- Dept. of Linguistics, University of Iowa, Iowa City, IA, USA
| | - Sarah Colby
- Dept. of Psychological & Brain Sciences, University of Iowa, Iowa City, IA, USA
- Dept. of Otolaryngology-Head and Neck Surgery, University of Iowa, Iowa City, IA, USA
| |
Collapse
|
7
|
Dor YI, Algom D, Shakuf V, Ben-David BM. Age-related differences in processing of emotions in speech disappear with babble noise in the background. Cogn Emot 2024:1-10. [PMID: 38764186 DOI: 10.1080/02699931.2024.2351960] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Accepted: 04/26/2024] [Indexed: 05/21/2024]
Abstract
Older adults process emotional speech differently than young adults, relying less on prosody (tone) relative to semantics (words). This study aimed to elucidate the mechanisms underlying these age-related differences via an emotional speech-in-noise test. A sample of 51 young and 47 older adults rated spoken sentences with emotional content on both prosody and semantics, presented on the background of wideband speech-spectrum noise (sensory interference) or on the background of multi-talker babble (sensory/cognitive interference). The presence of wideband noise eliminated age-related differences in semantics but not in prosody when processing emotional speech. Conversely, the presence of babble resulted in the elimination of age-related differences across all measures. The results suggest that both sensory and cognitive-linguistic factors contribute to age-related changes in emotional speech processing. Because real world conditions typically involve noisy background, our results highlight the importance of testing under such conditions.
Collapse
Affiliation(s)
- Yehuda I Dor
- School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Daniel Algom
- School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
- Department of Communication Disorders, Achva Academic College, Arugot, Israel
| | - Vered Shakuf
- Department of Communication Disorders, Achva Academic College, Arugot, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- KITE, Toronto Rehabilitation Institute, University Health Networks (UHN), Toronto, ON, Canada
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
8
|
Colby SE, McMurray B. Efficiency of spoken word recognition slows across the adult lifespan. Cognition 2023; 240:105588. [PMID: 37586157 PMCID: PMC10530619 DOI: 10.1016/j.cognition.2023.105588] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2022] [Revised: 07/26/2023] [Accepted: 08/03/2023] [Indexed: 08/18/2023]
Abstract
Spoken word recognition is a critical hub during language processing, linking hearing and perception to meaning and syntax. Words must be recognized quickly and efficiently as speech unfolds to be successfully integrated into conversation. This makes word recognition a computationally challenging process even for young, normal hearing adults. Older adults often experience declines in hearing and cognition, which could be linked by age-related declines in the cognitive processes specific to word recognition. However, it is unclear whether changes in word recognition across the lifespan can be accounted for by hearing or domain-general cognition. Participants (N = 107) responded to spoken words in a Visual World Paradigm task while their eyes were tracked to assess the real-time dynamics of word recognition. We examined several indices of word recognition from early adolescence through older adulthood (ages 11-78). The timing and proportion of eye fixations to target and competitor images reveals that spoken word recognition became more efficient through age 25 and began to slow in middle age, accompanied by declines in the ability to resolve competition (e.g., suppressing sandwich to recognize sandal). There was a unique effect of age even after accounting for differences in inhibitory control, processing speed, and hearing thresholds. This suggests a limited age range where listeners are peak performers.
Collapse
Affiliation(s)
- Sarah E Colby
- Department of Psychological and Brain Sciences, University of Iowa, Psychological and Brain Sciences Building, Iowa City, IA, 52242, USA; Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA, 52242, USA.
| | - Bob McMurray
- Department of Psychological and Brain Sciences, University of Iowa, Psychological and Brain Sciences Building, Iowa City, IA, 52242, USA; Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA, 52242, USA; Department of Communication Sciences and Disorders, University of Iowa, Wendell Johnson Speech and Hearing Center, Iowa City, IA, 52242, USA; Department of Linguistics, University of Iowa, Phillips Hall, Iowa City, IA 52242, USA
| |
Collapse
|
9
|
Hintz F, Voeten CC, Scharenborg O. Recognizing non-native spoken words in background noise increases interference from the native language. Psychon Bull Rev 2023; 30:1549-1563. [PMID: 36544064 PMCID: PMC10482792 DOI: 10.3758/s13423-022-02233-7] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/30/2022] [Indexed: 12/24/2022]
Abstract
Listeners frequently recognize spoken words in the presence of background noise. Previous research has shown that noise reduces phoneme intelligibility and hampers spoken-word recognition - especially for non-native listeners. In the present study, we investigated how noise influences lexical competition in both the non-native and the native language, reflecting the degree to which both languages are co-activated. We recorded the eye movements of native Dutch participants as they listened to English sentences containing a target word while looking at displays containing four objects. On target-present trials, the visual referent depicting the target word was present, along with three unrelated distractors. On target-absent trials, the target object (e.g., wizard) was absent. Instead, the display contained an English competitor, overlapping with the English target in phonological onset (e.g., window), a Dutch competitor, overlapping with the English target in phonological onset (e.g., wimpel, pennant), and two unrelated distractors. Half of the sentences was masked by speech-shaped noise; the other half was presented in quiet. Compared to speech in quiet, noise delayed fixations to the target objects on target-present trials. For target-absent trials, we observed that the likelihood for fixation biases towards the English and Dutch onset competitors (over the unrelated distractors) was larger in noise than in quiet. Our data thus show that the presence of background noise increases lexical competition in the task-relevant non-native (English) and in the task-irrelevant native (Dutch) language. The latter reflects stronger interference of one's native language during non-native spoken-word recognition under adverse conditions.
Collapse
Affiliation(s)
- Florian Hintz
- Max Planck Institute for Psycholinguistics, P.O. Box 310, 6500 AH, Nijmegen, The Netherlands.
| | | | - Odette Scharenborg
- Multimedia Computing Group, Delft University of Technology, Delft, Netherlands
| |
Collapse
|
10
|
Amos RM, Hartsuiker RJ, Seeber KG, Pickering MJ. Purposeful listening in challenging conditions: A study of prediction during consecutive interpreting in noise. PLoS One 2023; 18:e0288960. [PMID: 37471379 PMCID: PMC10359016 DOI: 10.1371/journal.pone.0288960] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2022] [Accepted: 07/10/2023] [Indexed: 07/22/2023] Open
Abstract
Prediction is often used during language comprehension. However, studies of prediction have tended to focus on L1 listeners in quiet conditions. Thus, it is unclear how listeners predict outside the laboratory and in specific communicative settings. Here, we report two eye-tracking studies which used a visual-world paradigm to investigate whether prediction during a consecutive interpreting task differs from prediction during a listening task in L2 listeners, and whether L2 listeners are able to predict in the noisy conditions that might be associated with this communicative setting. In a first study, thirty-six Dutch-English bilinguals either just listened to, or else listened to and then consecutively interpreted, predictable sentences presented on speech-shaped sound. In a second study, another thirty-six Dutch-English bilinguals carried out the same tasks in clear speech. Our results suggest that L2 listeners predict the meaning of upcoming words in noisy conditions. However, we did not find that predictive eye movements depended on task, nor that L2 listeners predicted upcoming word form. We also did not find a difference in predictive patterns when we compared our two studies. Thus, L2 listeners predict in noisy circumstances, supporting theories which posit that prediction regularly takes place in comprehension, but we did not find evidence that a subsequent production task or noise affects semantic prediction.
Collapse
Affiliation(s)
- Rhona M. Amos
- Department of Interpreting, Faculty of Translation and Interpreting, University of Geneva, Geneva, Switzerland
| | | | - Kilian G. Seeber
- Department of Interpreting, Faculty of Translation and Interpreting, University of Geneva, Geneva, Switzerland
| | - Martin J. Pickering
- Department of Psychology, School of Philosophy, Psychology and Language Sciences, University of Edinburgh, Edinburgh, Scotland
| |
Collapse
|
11
|
Klein KE, Walker EA, McMurray B. Delayed Lexical Access and Cascading Effects on Spreading Semantic Activation During Spoken Word Recognition in Children With Hearing Aids and Cochlear Implants: Evidence From Eye-Tracking. Ear Hear 2023; 44:338-357. [PMID: 36253909 PMCID: PMC9957808 DOI: 10.1097/aud.0000000000001286] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023]
Abstract
OBJECTIVE The objective of this study was to characterize the dynamics of real-time lexical access, including lexical competition among phonologically similar words, and spreading semantic activation in school-age children with hearing aids (HAs) and children with cochlear implants (CIs). We hypothesized that developing spoken language via degraded auditory input would lead children with HAs or CIs to adapt their approach to spoken word recognition, especially by slowing down lexical access. DESIGN Participants were children ages 9- to 12-years old with normal hearing (NH), HAs, or CIs. Participants completed a Visual World Paradigm task in which they heard a spoken word and selected the matching picture from four options. Competitor items were either phonologically similar, semantically similar, or unrelated to the target word. As the target word unfolded, children's fixations to the target word, cohort competitor, rhyme competitor, semantically related item, and unrelated item were recorded as indices of ongoing lexical access and spreading semantic activation. RESULTS Children with HAs and children with CIs showed slower fixations to the target, reduced fixations to the cohort competitor, and increased fixations to the rhyme competitor, relative to children with NH. This wait-and-see profile was more pronounced in the children with CIs than the children with HAs. Children with HAs and children with CIs also showed delayed fixations to the semantically related item, although this delay was attributable to their delay in activating words in general, not to a distinct semantic source. CONCLUSIONS Children with HAs and children with CIs showed qualitatively similar patterns of real-time spoken word recognition. Findings suggest that developing spoken language via degraded auditory input causes long-term cognitive adaptations to how listeners recognize spoken words, regardless of the type of hearing device used. Delayed lexical access directly led to delays in spreading semantic activation in children with HAs and CIs. This delay in semantic processing may impact these children's ability to understand connected speech in everyday life.
Collapse
Affiliation(s)
- Kelsey E Klein
- Department of Audiology and Speech Pathology, University of Tennessee Health Science Center, Knoxville, Tennessee, USA
| | - Elizabeth A Walker
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, Iowa, USA
| | - Bob McMurray
- Department of Psychological and Brain Sciences, Department of Communication Sciences and Disorders, and Department of Otolaryngology, University of Iowa, Iowa City, Iowa, USA
| |
Collapse
|
12
|
Lemel R, Shalev L, Nitsan G, Ben-David BM. Listen up! ADHD slows spoken-word processing in adverse listening conditions: Evidence from eye movements. RESEARCH IN DEVELOPMENTAL DISABILITIES 2023; 133:104401. [PMID: 36577332 DOI: 10.1016/j.ridd.2022.104401] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/23/2022] [Revised: 10/23/2022] [Accepted: 12/16/2022] [Indexed: 06/17/2023]
Abstract
BACKGROUND Cognitive skills such as sustained attention, inhibition and working memory are essential for speech processing, yet are often impaired in people with ADHD. Offline measures have indicated difficulties in speech recognition on multi-talker babble (MTB) background for young adults with ADHD (yaADHD). However, to-date no study has directly tested online speech processing in adverse conditions for yaADHD. AIMS Gauging the effects of ADHD on segregating the spoken target-word from its sound-sharing competitor, in MTB and working-memory (WM) load. METHODS AND PROCEDURES Twenty-four yaADHD and 22 matched controls that differ in sustained attention (SA) but not in WM were asked to follow spoken instructions presented on MTB to touch a named object, while retaining one (low-load) or four (high-load) digit/s for later recall. Their eye fixations were tracked. OUTCOMES AND RESULTS In the high-load condition, speech processing was less accurate and slowed by 140ms for yaADHD. In the low-load condition, the processing advantage shifted from early perceptual to later cognitive stages. Fixation transitions (hesitations) were inflated for yaADHD. CONCLUSIONS AND IMPLICATIONS ADHD slows speech processing in adverse listening conditions and increases hesitation, as speech unfolds in time. These effects, detected only by online eyetracking, relate to attentional difficulties. We suggest online speech processing as a novel purview on ADHD. WHAT THIS PAPER ADDS?: We suggest speech processing in adverse listening conditions as a novel vantage point on ADHD. Successful speech recognition in noise is essential for performance across daily settings: academic, employment and social interactions. It involves several executive functions, such as inhibition and sustained attention. Impaired performance in these functions is characteristic of ADHD. However, to date there is only scant research on speech processing in ADHD. The current study is the first to investigate online speech processing as the word unfolds in time using eyetracking for young adults with ADHD (yaADHD). This method uncovered slower speech processing in multi-talker babble noise for yaADHD compared to matched controls. The performance of yaADHD indicated increased hesitation between the spoken word and sound-sharing alternatives (e.g., CANdle-CANdy). These delays and hesitations, on the single word level, could accumulate in continuous speech to significantly impair communication in ADHD, with severe implications on their quality of life and academic success. Interestingly, whereas yaADHD and controls were matched on WM standardized tests, WM load appears to affect speech processing for yaADHD more than for controls. This suggests that ADHD may lead to inefficient deployment of WM resources that may not be detected when WM is tested alone. Note that these intricate differences could not be detected using traditional offline accuracy measures, further supporting the use of eyetracking in speech tasks. Finally, communication is vital for active living and wellbeing. We suggest paying attention to speech processing in ADHD in treatment and when considering accessibility and inclusion.
Collapse
Affiliation(s)
- Rony Lemel
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Lilach Shalev
- Constantiner School of Education and Sagol School of Neuroscience, Tel-Aviv University, Tel-Aviv, Israel
| | - Gal Nitsan
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel; Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel; Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada; Toronto Rehabilitation Institute, University Health Networks (UHN), ON, Canada.
| |
Collapse
|
13
|
McMurray B. I'm not sure that curve means what you think it means: Toward a [more] realistic understanding of the role of eye-movement generation in the Visual World Paradigm. Psychon Bull Rev 2023; 30:102-146. [PMID: 35962241 PMCID: PMC10964151 DOI: 10.3758/s13423-022-02143-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 06/29/2022] [Indexed: 11/08/2022]
Abstract
The Visual World Paradigm (VWP) is a powerful experimental paradigm for language research. Listeners respond to speech in a "visual world" containing potential referents of the speech. Fixations to these referents provides insight into the preliminary states of language processing as decisions unfold. The VWP has become the dominant paradigm in psycholinguistics and extended to every level of language, development, and disorders. Part of its impact is the impressive data visualizations which reveal the millisecond-by-millisecond time course of processing, and advances have been made in developing new analyses that precisely characterize this time course. All theoretical and statistical approaches make the tacit assumption that the time course of fixations is closely related to the underlying activation in the system. However, given the serial nature of fixations and their long refractory period, it is unclear how closely the observed dynamics of the fixation curves are actually coupled to the underlying dynamics of activation. I investigated this assumption with a series of simulations. Each simulation starts with a set of true underlying activation functions and generates simulated fixations using a simple stochastic sampling procedure that respects the sequential nature of fixations. I then analyzed the results to determine the conditions under which the observed fixations curves match the underlying functions, the reliability of the observed data, and the implications for Type I error and power. These simulations demonstrate that even under the simplest fixation-based models, observed fixation curves are systematically biased relative to the underlying activation functions, and they are substantially noisier, with important implications for reliability and power. I then present a potential generative model that may ultimately overcome many of these issues.
Collapse
Affiliation(s)
- Bob McMurray
- Department of Psychological and Brain Sciences, 278 PBSB, University of Iowa, Iowa City, IA, 52242, USA.
- Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA, USA.
- Department of Linguistics, University of Iowa, Iowa City, IA, USA.
- Department of Otolaryngology, University of Iowa, Iowa City, IA, USA.
| |
Collapse
|
14
|
Liu F, Jiang S, Kang J, Wu Y, Yang D, Meng Q, Wang C. On the definition of noise. HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS 2022; 9:406. [PMID: 36406149 PMCID: PMC9643889 DOI: 10.1057/s41599-022-01431-x] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/30/2021] [Accepted: 10/27/2022] [Indexed: 06/16/2023]
Abstract
Urbanization has exposed people to extreme sound levels. Although researchers have investigated the ability of people to listen, analyze, and distinguish sound, the concept of noise has not been clearly articulated from a human perspective. The lack of knowledge on how people perceive noise limits our capacity to control it in a targeted manner. This study aimed to interpret the definition of noise from the public perspective based on a grounded theory approach. Seventy-eight participants were interviewed about noise, and four categories of perceived understanding of noise were identified: challenges, definitions of noise, opportunities, and action. As one of the challenges, urbanization is associated with increased noise levels around the human environment. In terms of definition, perceiving sound as noise is considered to be a result of the complex and dynamic process that includes sound, the environment, and humans. Sound and humans interact with the environment. In terms of opportunities, noise may have positive roles on certain occasions, dispelling the misconception that noise is exclusively negative. In addition, we found that noise perception has gradually shifted from noise control to noise utilization. In terms of action, noise can be controlled at the sound sources, susceptible target groups, susceptible behaviors and states, locations, and times where noise is perceived with high frequency. In this study, we investigated several aspects of noise, ranging from noise control, soundscape definition, and 'soundscape indices' (SSID) integration and application. Our findings provide an additional basis for developing better definitions, control, and utilization strategies of noise in the future, thereby improving the quality of the sound environment.
Collapse
Affiliation(s)
- Fangfang Liu
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| | - Shan Jiang
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| | - Jian Kang
- Institute for Environmental Design and Engineering, University College London, London, WC1H 0NN UK
| | - Yue Wu
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| | - Da Yang
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| | - Qi Meng
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| | - Chaowei Wang
- Heilongjiang Cold Region Architectural Science Key Laboratory, School of Architecture, Harbin Institute of Technology, 66 West Dazhi Street, Nan Gang District, 150006 Harbin, PR China
| |
Collapse
|
15
|
McMurray B, Sarrett ME, Chiu S, Black AK, Wang A, Canale R, Aslin RN. Decoding the temporal dynamics of spoken word and nonword processing from EEG. Neuroimage 2022; 260:119457. [PMID: 35842096 PMCID: PMC10875705 DOI: 10.1016/j.neuroimage.2022.119457] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2022] [Revised: 07/02/2022] [Accepted: 07/06/2022] [Indexed: 11/23/2022] Open
Abstract
The efficiency of spoken word recognition is essential for real-time communication. There is consensus that this efficiency relies on an implicit process of activating multiple word candidates that compete for recognition as the acoustic signal unfolds in real-time. However, few methods capture the neural basis of this dynamic competition on a msec-by-msec basis. This is crucial for understanding the neuroscience of language, and for understanding hearing, language and cognitive disorders in people for whom current behavioral methods are not suitable. We applied machine-learning techniques to standard EEG signals to decode which word was heard on each trial and analyzed the patterns of confusion over time. Results mirrored psycholinguistic findings: Early on, the decoder was equally likely to report the target (e.g., baggage) or a similar sounding competitor (badger), but by around 500 msec, competitors were suppressed. Follow up analyses show that this is robust across EEG systems (gel and saline), with fewer channels, and with fewer trials. Results are robust within individuals and show high reliability. This suggests a powerful and simple paradigm that can assess the neural dynamics of speech decoding, with potential applications for understanding lexical development in a variety of clinical disorders.
Collapse
Affiliation(s)
- Bob McMurray
- Dept. of Psychological and Brain Sciences, Dept. of Communication Sciences and Disorders, Dept. of Linguistics and Dept. of Otolaryngology, University of Iowa.
| | - McCall E Sarrett
- Interdisciplinary Graduate Program in Neuroscience, Unviersity of Iowa
| | - Samantha Chiu
- Dept. of Psychological and Brain Sciences, University of Iowa
| | - Alexis K Black
- School of Audiology and Speech Sciences, University of British Columbia, Haskins Laboratories
| | - Alice Wang
- Dept. of Psychology, University of Oregon, Haskins Laboratories
| | - Rebecca Canale
- Dept. of Psychological Sciences, University of Connecticut, Haskins Laboratories
| | - Richard N Aslin
- Haskins Laboratories, Department of Psychology and Child Study Center, Yale University, Department of Psychology, University of Connecticut
| |
Collapse
|
16
|
Smith FX, McMurray B. Lexical Access Changes Based on Listener Needs: Real-Time Word Recognition in Continuous Speech in Cochlear Implant Users. Ear Hear 2022; 43:1487-1501. [PMID: 35067570 PMCID: PMC9300769 DOI: 10.1097/aud.0000000000001203] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
Abstract
OBJECTIVES A key challenge in word recognition is the temporary ambiguity created by the fact that speech unfolds over time. In normal hearing (NH) listeners, this temporary ambiguity is resolved through incremental processing and competition among lexical candidates. Post-lingually deafened cochlear implant (CI) users show similar incremental processing and competition but with slight delays. However, even brief delays could lead to drastic changes when compounded across multiple words in a phrase. This study asks whether words presented in non-informative continuous speech (a carrier phrase) are processed differently than in isolation and whether NH listeners and CI users exhibit different effects of a carrier phrase. DESIGN In a Visual World Paradigm experiment, listeners heard words either in isolation or in non-informative carrier phrases (e.g., "click on the…" ). Listeners selected the picture corresponding to the target word from among four items including the target word (e.g., mustard ), a cohort competitor (e.g., mustache ), a rhyme competitor (e.g., custard ), and an unrelated item (e.g., penguin ). Eye movements were tracked as an index of the relative activation of each lexical candidate as competition unfolds over the course of word recognition. Participants included 21 post-lingually deafened cochlear implant users and 21 NH controls. A replication experiment presented in the Supplemental Digital Content, http://links.lww.com/EANDH/A999 included an additional 22 post-lingually deafened CI users and 18 NH controls. RESULTS Both CI users and the NH controls were accurate at recognizing the words both in continuous speech and in isolation. The time course of lexical activation (indexed by the fixations) differed substantially between groups. CI users were delayed in fixating the target relative to NH controls. Additionally, CI users showed less competition from cohorts than NH controls (even as previous studies have often report increased competition). However, CI users took longer to suppress the cohort and suppressed it less fully than the NH controls. For both CI users and NH controls, embedding words in carrier phrases led to more immediacy in lexical access as observed by increases in cohort competition relative to when words were presented in isolation. However, CI users were not differentially affected by the carriers. CONCLUSIONS Unlike prior work, CI users appeared to exhibit "wait-and-see" profile, in which lexical access is delayed minimizing early competition. However, CI users simultaneously sustained competitor activation late in the trial, possibly to preserve flexibility. This hybrid profile has not been observed previously. When target words are heard in continuous speech, both CI users and NH controls more heavily weight early information. However, CI users (but not NH listeners) also commit less fully to the target, potentially keeping options open if they need to recover from a misperception. This mix of patterns reflects a lexical system that is extremely flexible and adapts to fit the needs of a listener.
Collapse
Affiliation(s)
| | - Bob McMurray
- Dept. of Psychological and Brain Sciences, University of Iowa
- Dept. of Otolaryngology, University of Iowa
| |
Collapse
|
17
|
Nitsan G, Baharav S, Tal-Shir D, Shakuf V, Ben-David BM. Speech Processing as a Far-Transfer Gauge of Serious Games for Cognitive Training in Aging: Randomized Controlled Trial of Web-Based Effectivate Training. JMIR Serious Games 2022; 10:e32297. [PMID: 35900825 PMCID: PMC9400949 DOI: 10.2196/32297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2021] [Revised: 04/21/2022] [Accepted: 04/28/2022] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND The number of serious games for cognitive training in aging (SGCTAs) is proliferating in the market and attempting to combat one of the most feared aspects of aging-cognitive decline. However, the efficacy of many SGCTAs is still questionable. Even the measures used to validate SGCTAs are up for debate, with most studies using cognitive measures that gauge improvement in trained tasks, also known as near transfer. This study takes a different approach, testing the efficacy of the SGCTA-Effectivate-in generating tangible far-transfer improvements in a nontrained task-the Eye tracking of Word Identification in Noise Under Memory Increased Load (E-WINDMIL)-which tests speech processing in adverse conditions. OBJECTIVE This study aimed to validate the use of a real-time measure of speech processing as a gauge of the far-transfer efficacy of an SGCTA designed to train executive functions. METHODS In a randomized controlled trial that included 40 participants, we tested 20 (50%) older adults before and after self-administering the SGCTA Effectivate training and compared their performance with that of the control group of 20 (50%) older adults. The E-WINDMIL eye-tracking task was administered to all participants by blinded experimenters in 2 sessions separated by 2 to 8 weeks. RESULTS Specifically, we tested the change between sessions in the efficiency of segregating the spoken target word from its sound-sharing alternative, as the word unfolds in time. We found that training with the SGCTA Effectivate improved both early and late speech processing in adverse conditions, with higher discrimination scores in the training group than in the control group (early processing: F1,38=7.371; P=.01; ηp2=0.162 and late processing: F1,38=9.003; P=.005; ηp2=0.192). CONCLUSIONS This study found the E-WINDMIL measure of speech processing to be a valid gauge for the far-transfer effects of executive function training. As the SGCTA Effectivate does not train any auditory task or language processing, our results provide preliminary support for the ability of Effectivate to create a generalized cognitive improvement. Given the crucial role of speech processing in healthy and successful aging, we encourage researchers and developers to use speech processing measures, the E-WINDMIL in particular, to gauge the efficacy of SGCTAs. We advocate for increased industry-wide adoption of far-transfer metrics to gauge SGCTAs.
Collapse
Affiliation(s)
- Gal Nitsan
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel.,Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Shai Baharav
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Dalith Tal-Shir
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Vered Shakuf
- Department of Communications Disorders, Achva Academic College, Arugot, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel.,Toronto Rehabilitation Institute, University Health Networks, Toronto, ON, Canada.,Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
18
|
Saryazdi R, Nuque J, Chambers CG. Pragmatic inferences in aging and human-robot communication. Cognition 2022; 223:105017. [PMID: 35131577 DOI: 10.1016/j.cognition.2022.105017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2020] [Revised: 06/12/2021] [Accepted: 01/05/2022] [Indexed: 12/30/2022]
Abstract
Despite the increase in research on older adults' communicative behavior, little work has explored patterns of age-related change in pragmatic inferencing and how these patterns are adapted depending on the situation-specific context. In two eye-tracking experiments, participants followed instructions like "Click on the greenhouse", which were either played over speakers or spoken live by a co-present robot partner. Implicit inferential processes were measured by exploring the extent to which listeners temporarily (mis)understood the unfolding noun to be a modified phrase referring to a competitor object in the display (green hat). This competitor was accompanied by either another member of the same category or an unrelated item (tan hat vs. dice). Experiment 1 (no robot) showed clear evidence of contrastive inferencing in both younger and older adults (more looks to the green hat when the tan hat was also present). Experiment 2 explored the ability to suppress these contrastive inferences when the robot talker was known to lack any color perception, making descriptions like "green hat" implausible. Younger but not older listeners were able to suppress contrastive inferences in this context, suggesting older adults could not keep the relevant limitations in mind and/or were more likely to spontaneously ascribe human attributes to the robot. Together, the findings enhance our understanding of pragmatic inferencing in aging.
Collapse
Affiliation(s)
- Raheleh Saryazdi
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada; Department of Psychology, University of Toronto, Mississauga, Ontario, Canada.
| | - Joanne Nuque
- Department of Psychology, University of Toronto, Mississauga, Ontario, Canada
| | - Craig G Chambers
- Department of Psychology, University of Toronto, Mississauga, Ontario, Canada
| |
Collapse
|
19
|
Dor YI, Algom D, Shakuf V, Ben-David BM. Age-Related Changes in the Perception of Emotions in Speech: Assessing Thresholds of Prosody and Semantics Recognition in Noise for Young and Older Adults. Front Neurosci 2022; 16:846117. [PMID: 35546888 PMCID: PMC9082150 DOI: 10.3389/fnins.2022.846117] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2021] [Accepted: 03/14/2022] [Indexed: 11/15/2022] Open
Abstract
Older adults process emotions in speech differently than do young adults. However, it is unclear whether these age-related changes impact all speech channels to the same extent, and whether they originate from a sensory or a cognitive source. The current study adopted a psychophysical approach to directly compare young and older adults’ sensory thresholds for emotion recognition in two channels of spoken-emotions: prosody (tone) and semantics (words). A total of 29 young adults and 26 older adults listened to 50 spoken sentences presenting different combinations of emotions across prosody and semantics. They were asked to recognize the prosodic or semantic emotion, in separate tasks. Sentences were presented on the background of speech-spectrum noise ranging from SNR of −15 dB (difficult) to +5 dB (easy). Individual recognition thresholds were calculated (by fitting psychometric functions) separately for prosodic and semantic recognition. Results indicated that: (1). recognition thresholds were better for young over older adults, suggesting an age-related general decrease across channels; (2). recognition thresholds were better for prosody over semantics, suggesting a prosodic advantage; (3). importantly, the prosodic advantage in thresholds did not differ between age groups (thus a sensory source for age-related differences in spoken-emotions processing was not supported); and (4). larger failures of selective attention were found for older adults than for young adults, indicating that older adults experienced larger difficulties in inhibiting irrelevant information. Taken together, results do not support a sole sensory source, but rather an interplay of cognitive and sensory sources for age-related differences in spoken-emotions processing.
Collapse
Affiliation(s)
- Yehuda I Dor
- School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel.,Communication, Aging and Neuropsychology Lab (CANlab), Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
| | - Daniel Algom
- School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
| | - Vered Shakuf
- Department of Communications Disorders, Achva Academic College, Arugot, Israel
| | - Boaz M Ben-David
- Communication, Aging and Neuropsychology Lab (CANlab), Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel.,Toronto Rehabilitation Institute, University Health Networks (UHN), Toronto, ON, Canada.,Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
| |
Collapse
|
20
|
Children’s and adults’ use of fictional discourse and semantic knowledge for prediction in language processing. PLoS One 2022; 17:e0267297. [PMID: 35482807 PMCID: PMC9049568 DOI: 10.1371/journal.pone.0267297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Accepted: 04/06/2022] [Indexed: 11/20/2022] Open
Abstract
Using real-time eye-movement measures, we asked how a fantastical discourse context competes with stored representations of real-world events to influence the moment-by-moment interpretation of a story by 7-year-old children and adults. Seven-year-olds were less effective at bypassing stored real-world knowledge during real-time interpretation than adults. Our results suggest that children privilege stored semantic knowledge over situation-specific information presented in a fictional story context. We suggest that 7-year-olds’ canonical semantic and conceptual relations are sufficiently strongly rooted in statistical patterns in language that have consolidated over time that they overwhelm new and unexpected information even when the latter is fantastical and highly salient.
Collapse
|
21
|
Nitsan G, Banai K, Ben-David BM. One Size Does Not Fit All: Examining the Effects of Working Memory Capacity on Spoken Word Recognition in Older Adults Using Eye Tracking. Front Psychol 2022; 13:841466. [PMID: 35478743 PMCID: PMC9037998 DOI: 10.3389/fpsyg.2022.841466] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2021] [Accepted: 03/14/2022] [Indexed: 11/13/2022] Open
Abstract
Difficulties understanding speech form one of the most prevalent complaints among older adults. Successful speech perception depends on top-down linguistic and cognitive processes that interact with the bottom-up sensory processing of the incoming acoustic information. The relative roles of these processes in age-related difficulties in speech perception, especially when listening conditions are not ideal, are still unclear. In the current study, we asked whether older adults with a larger working memory capacity process speech more efficiently than peers with lower capacity when speech is presented in noise, with another task performed in tandem. Using the Eye-tracking of Word Identification in Noise Under Memory Increased Load (E-WINDMIL) an adapted version of the "visual world" paradigm, 36 older listeners were asked to follow spoken instructions presented in background noise, while retaining digits for later recall under low (single-digit) or high (four-digits) memory load. In critical trials, instructions (e.g., "point at the candle") directed listeners' gaze to pictures of objects whose names shared onset or offset sounds with the name of a competitor that was displayed on the screen at the same time (e.g., candy or sandal). We compared listeners with different memory capacities on the time course for spoken word recognition under the two memory loads by testing eye-fixations on a named object, relative to fixations on an object whose name shared phonology with the named object. Results indicated two trends. (1) For older adults with lower working memory capacity, increased memory load did not affect online speech processing, however, it impaired offline word recognition accuracy. (2) The reverse pattern was observed for older adults with higher working memory capacity: increased task difficulty significantly decreases online speech processing efficiency but had no effect on offline word recognition accuracy. Results suggest that in older adults, adaptation to adverse listening conditions is at least partially supported by cognitive reserve. Therefore, additional cognitive capacity may lead to greater resilience of older listeners to adverse listening conditions. The differential effects documented by eye movements and accuracy highlight the importance of using both online and offline measures of speech processing to explore age-related changes in speech perception.
Collapse
Affiliation(s)
- Gal Nitsan
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
| | - Karen Banai
- Department of Communication Sciences and Disorders, University of Haifa, Haifa, Israel
| | - Boaz M. Ben-David
- Baruch Ivcher School of Psychology, Reichman University (IDC), Herzliya, Israel
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
- Toronto Rehabilitation Institute, University Health Networks, Toronto, ON, Canada
| |
Collapse
|
22
|
Saryazdi R, Nuque J, Chambers CG. Linguistic Redundancy and its Effects on Younger and Older Adults' Real-Time Comprehension and Memory. Cogn Sci 2022; 46:e13123. [PMID: 35377508 DOI: 10.1111/cogs.13123] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2021] [Revised: 01/05/2022] [Accepted: 02/14/2022] [Indexed: 01/12/2023]
Abstract
Redundant modifiers can facilitate referential interpretation by narrowing attention to intended referents. This is intriguing because, on traditional accounts, redundancy should impair comprehension. Little is known, however, about the effects of redundancy on older adults' comprehension. Older adults may show different patterns due to age-related decline (e.g., processing speed and memory) or their greater proclivity for linguistic redundancy, as suggested in language production studies. The present study explores the effects of linguistic redundancy on younger and older listeners' incremental referential processing, judgments of informativity, and downstream memory performance. In an eye tracking task, gaze was monitored as listeners followed instructions from a social robot referring to a unique object within a multi-object display. Critical trials were varied in terms of modifier type ("…closed/purple/[NONE] umbrella") and whether displays contained another object matching target properties (closed purple notebook), making modifiers less effective at narrowing attention. Relative to unmodified descriptions, redundant color modifiers facilitated comprehension, particularly when they narrowed attention to a single referent. Descriptions with redundant state modifiers always impaired real-time comprehension. In contrast, memory measures showed faster recognition of objects previously described with redundant state modifiers. Although color and state descriptions had different effects on referential processing and memory, informativity judgments showed participants perceived them as informationally redundant to the same extent relative to unmodified descriptions. Importantly, the patterns did not differ by listener age. Together, the results show that the effects of linguistic redundancy are stable across adulthood but vary as a function of modifier type, visual context, and the measured phenomenon.
Collapse
Affiliation(s)
- Raheleh Saryazdi
- Department of Psychology, University of Toronto.,Department of Psychology, University of Toronto Mississauga
| | - Joanne Nuque
- Department of Psychology, University of Toronto Mississauga
| | - Craig G Chambers
- Department of Psychology, University of Toronto.,Department of Psychology, University of Toronto Mississauga
| |
Collapse
|
23
|
Braza MD, Porter HL, Buss E, Calandruccio L, McCreery RW, Leibold LJ. Effects of word familiarity and receptive vocabulary size on speech-in-noise recognition among young adults with normal hearing. PLoS One 2022; 17:e0264581. [PMID: 35271608 PMCID: PMC8912124 DOI: 10.1371/journal.pone.0264581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Accepted: 02/11/2022] [Indexed: 11/29/2022] Open
Abstract
Having a large receptive vocabulary benefits speech-in-noise recognition for young children, though this is not always the case for older children or adults. These observations could indicate that effects of receptive vocabulary size on speech-in-noise recognition differ depending on familiarity of the target words, with effects observed only for more recently acquired and less frequent words. Two experiments were conducted to evaluate effects of vocabulary size on open-set speech-in-noise recognition for adults with normal hearing. Targets were words acquired at 4, 9, 12 and 15 years of age, and they were presented at signal-to-noise ratios (SNRs) of -5 and -7 dB. Percent correct scores tended to fall with increasing age of acquisition (AoA), with the caveat that performance at -7 dB SNR was better for words acquired at 9 years of age than earlier- or later-acquired words. Similar results were obtained whether the AoA of the target words was blocked or mixed across trials. Differences in word duration appear to account for nonmonotonic effects of AoA. For all conditions, a positive correlation was observed between recognition and vocabulary size irrespective of target word AoA, indicating that effects of vocabulary size are not limited to recently acquired words. This dataset does not support differential assessment of AoA, lexical frequency, and other stimulus features known to affect lexical access.
Collapse
Affiliation(s)
- Meredith D. Braza
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, United States of America
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, United States of America
| | - Heather L. Porter
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, United States of America
| | - Emily Buss
- Department of Otolaryngology/Head and Neck Surgery, The University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, United States of America
| | - Lauren Calandruccio
- Department of Psychological Sciences, Case Western Reserve University, Cleveland, Ohio, United States of America
| | - Ryan W. McCreery
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, United States of America
| | - Lori J. Leibold
- Center for Hearing Research, Boys Town National Research Hospital, Omaha, Nebraska, United States of America
| |
Collapse
|
24
|
Keisari S, Feniger-Schaal R, Palgi Y, Golland Y, Gesser-Edelsburg A, Ben-David B. Synchrony in Old Age: Playing the Mirror Game Improves Cognitive Performance. Clin Gerontol 2022; 45:312-326. [PMID: 32762289 DOI: 10.1080/07317115.2020.1799131] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/21/2023]
Abstract
OBJECTIVES Studies have shown that synchronized motion between people positively affects a range of emotional and social functions. The mirror-game is a synchrony-based paradigm, common to theater, performance arts, and therapy, which includes dyadic synchronized motion, playfulness, and spontaneity. The goal of the current study is to examine the effects of the mirror-game on subjective and cognitive indices in late life. METHODS Thirty-four older adults (aged 71-98) participated in a within-group study design. Participants conducted two sessions of 9-minute movement activities: the mirror-game and the control condition - a physical exercise class. Several measures were taken before and after experimental sessions to assess socio-emotional and attentional functions. RESULTS The mirror-game enhanced performance on the attention sub-scale and led to faster detections of spoken words in noise. Further, it enhanced perceived partner responsiveness and led to an increase in positive reported experience. CONCLUSIONS Our preliminary findings suggest that the mirror-game, rather than the exercise class, may have an immediate impact on mood and some attentional functions. CLINICAL IMPLICATIONS The mirror-game is a novel intervention, with potential benefits of social-emotional and cognitive functioning, which can be easily implemented into the daily routine care of older adults. Future studies should explore the effect of the mirror-game on additional cognitive and socio-emotional aspects.
Collapse
Affiliation(s)
- Shoshi Keisari
- Department of Gerontology, University of Haifa, Haifa, Israel
| | - Rinat Feniger-Schaal
- School of Creative Arts Therapies, The Center for the Study of Child Development, University of Haifa, Haifa, Israel
| | - Yuval Palgi
- Department of Gerontology, University of Haifa, Haifa, Israel.,The Center for Research and Study of Aging, University of Haifa, Haifa, Israel
| | - Yulia Golland
- Sagol Center for Brain and Mind, Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC) Herzliya, Herzliya, Israel
| | - Anat Gesser-Edelsburg
- School of Public Health and the Health and Risk Communication Research Center, University of Haifa, Haifa, Israel
| | - Boaz Ben-David
- Baruch Ivcher School of Psychology, The Interdisciplinary Center (IDC) Herzliya, Herzliya, Israel.,Department of Speech-language Pathology, University of Toronto, Toronto, Canada
| |
Collapse
|
25
|
Karaminis T, Hintz F, Scharenborg O. The Presence of Background Noise Extends the Competitor Space in Native and Non-Native Spoken-Word Recognition: Insights from Computational Modeling. Cogn Sci 2022; 46:e13110. [PMID: 35188686 PMCID: PMC9286693 DOI: 10.1111/cogs.13110] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Revised: 12/17/2021] [Accepted: 01/23/2022] [Indexed: 11/29/2022]
Abstract
Oral communication often takes place in noisy environments, which challenge spoken‐word recognition. Previous research has suggested that the presence of background noise extends the number of candidate words competing with the target word for recognition and that this extension affects the time course and accuracy of spoken‐word recognition. In this study, we further investigated the temporal dynamics of competition processes in the presence of background noise, and how these vary in listeners with different language proficiency (i.e., native and non‐native) using computational modeling. We developed ListenIN (Listen‐In‐Noise), a neural‐network model based on an autoencoder architecture, which learns to map phonological forms onto meanings in two languages and simulates native and non‐native spoken‐word comprehension. We also examined the model's activation states during online spoken‐word recognition. These analyses demonstrated that the presence of background noise increases the number of competitor words, which are engaged in phonological competition and that this happens in similar ways intra and interlinguistically and in native and non‐native listening. Taken together, our results support accounts positing a “many‐additional‐competitors scenario” for the effects of noise on spoken‐word recognition.
Collapse
Affiliation(s)
| | - Florian Hintz
- Department of Psychology of Language, Max Planck Institute for Psycholinguistics
| | | |
Collapse
|
26
|
More Than Words: the Relative Roles of Prosody and Semantics in the Perception of Emotions in Spoken Language by Postlingual Cochlear Implant Users. Ear Hear 2022; 43:1378-1389. [PMID: 35030551 DOI: 10.1097/aud.0000000000001199] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
OBJECTIVES The processing of emotional speech calls for the perception and integration of semantic and prosodic cues. Although cochlear implants allow for significant auditory improvements, they are limited in the transmission of spectro-temporal fine-structure information that may not support the processing of voice pitch cues. The goal of the current study is to compare the performance of postlingual cochlear implant (CI) users and a matched control group on perception, selective attention, and integration of emotional semantics and prosody. DESIGN Fifteen CI users and 15 normal hearing (NH) peers (age range, 18-65 years) 1istened to spoken sentences composed of different combinations of four discrete emotions (anger, happiness, sadness, and neutrality) presented in prosodic and semantic channels-T-RES: Test for Rating Emotions in Speech. In three separate tasks, listeners were asked to attend to the sentence as a whole, thus integrating both speech channels (integration), or to focus on one channel only (rating of target emotion) and ignore the other (selective attention). Their task was to rate how much they agreed that the sentence conveyed each of the predefined emotions. In addition, all participants performed standard tests of speech perception. RESULTS When asked to focus on one channel, semantics or prosody, both groups rated emotions similarly with comparable levels of selective attention. When the task was called for channel integration, group differences were found. CI users appeared to use semantic emotional information more than did their NH peers. CI users assigned higher ratings than did their NH peers to sentences that did not present the target emotion, indicating some degree of confusion. In addition, for CI users, individual differences in speech comprehension over the phone and identification of intonation were significantly related to emotional semantic and prosodic ratings, respectively. CONCLUSIONS CI users and NH controls did not differ in perception of prosodic and semantic emotions and in auditory selective attention. However, when the task called for integration of prosody and semantics, CI users overused the semantic information (as compared with NH). We suggest that as CI users adopt diverse cue weighting strategies with device experience, their weighting of prosody and semantics differs from those used by NH. Finally, CI users may benefit from rehabilitation strategies that strengthen perception of prosodic information to better understand emotional speech.
Collapse
|
27
|
Feng Y, Chen F. Nonintrusive objective measurement of speech intelligibility: A review of methodology. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2021.103204] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
28
|
Effect of Noise Reduction on Cortical Speech-in-Noise Processing and Its Variance due to Individual Noise Tolerance. Ear Hear 2021; 43:849-861. [PMID: 34751679 PMCID: PMC9010348 DOI: 10.1097/aud.0000000000001144] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Despite the widespread use of noise reduction (NR) in modern digital hearing aids, our neurophysiological understanding of how NR affects speech-in-noise perception and why its effect is variable is limited. The current study aimed to (1) characterize the effect of NR on the neural processing of target speech and (2) seek neural determinants of individual differences in the NR effect on speech-in-noise performance, hypothesizing that an individual's own capability to inhibit background noise would inversely predict NR benefits in speech-in-noise perception. DESIGN Thirty-six adult listeners with normal hearing participated in the study. Behavioral and electroencephalographic responses were simultaneously obtained during a speech-in-noise task in which natural monosyllabic words were presented at three different signal-to-noise ratios, each with NR off and on. A within-subject analysis assessed the effect of NR on cortical evoked responses to target speech in the temporal-frontal speech and language brain regions, including supramarginal gyrus and inferior frontal gyrus in the left hemisphere. In addition, an across-subject analysis related an individual's tolerance to noise, measured as the amplitude ratio of auditory-cortical responses to target speech and background noise, to their speech-in-noise performance. RESULTS At the group level, in the poorest signal-to-noise ratio condition, NR significantly increased early supramarginal gyrus activity and decreased late inferior frontal gyrus activity, indicating a switch to more immediate lexical access and less effortful cognitive processing, although no improvement in behavioral performance was found. The across-subject analysis revealed that the cortical index of individual noise tolerance significantly correlated with NR-driven changes in speech-in-noise performance. CONCLUSIONS NR can facilitate speech-in-noise processing despite no improvement in behavioral performance. Findings from the current study also indicate that people with lower noise tolerance are more likely to get more benefits from NR. Overall, results suggest that future research should take a mechanistic approach to NR outcomes and individual noise tolerance.
Collapse
|
29
|
Listening Effort Measured Using a Dual-task Paradigm in Adults With Different Amounts of Noise Exposure. Ear Hear 2021; 43:899-912. [PMID: 34619684 DOI: 10.1097/aud.0000000000001138] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Excessive noise levels can result in hearing damage and/or hearing-related symptoms, thereby leading to impaired communication and, eventually a decrease in the quality of life. Moreover, in daily practice, subjects often indicate that listening in noisy situations is a challenging and often exhausting experience, even in the absence of a clinically significant hearing loss. Hence, a person's perceived difficulty of the listening situation will also be important to consider. It has been suggested that beyond the peripheral factors, there are some central cognitive correlates of speech understanding that are essential for communication abilities. The aim of the present study was to evaluate the effect of the amount of noise exposure on hearing as measured by pure-tone audiometry and otoacoustic emissions (OAEs) on the one hand and listening effort measured using a dual-task paradigm on the other hand. DESIGN The study included 152 adults between 18 and 40 years of age. First, participants completed a self-administered questionnaire regarding the amount of noise exposure. Second, hearing status was evaluated using pure-tone audiometry and transient evoked OAEs (TEOAEs) as well as distortion product OAEs (DPOAEs). Finally, listening effort was evaluated using a dual-task paradigm, which consisted of a primary speech-recognition task in different listening conditions and a secondary visual memory task that were performed both separately and simultaneously. Based on the quartiles of their subjective estimation of noise exposure, the participants were categorized into a group with low noise exposure (lower quartile), moderate noise exposure (two middle quartiles), and high noise exposure (upper quartile). RESULTS There was no significant difference in hearing thresholds between the participants with low, moderate, and high noise exposure for each frequency of the pure-tone audiometry. In contrast, more absent TEOAEs and DPOAEs were found in the high noise exposed group. Regarding the primary speech-recognition task of the dual-task paradigm, no significant difference in speech recognition was found between the different groups of noise exposure. For the amount of listening effort, it was found that across all listening conditions subjects with high noise exposure expend significantly more listening effort compared with subjects with low and moderate noise exposure. CONCLUSIONS This study is a first exploration of the effects of different amounts of noise exposure on listening effort showing that, listening effort is increased in subjects with high noise exposure compared with subjects with low and medium noise exposure. The most plausible hypothesis pertains to an effect of noise exposure on the peripheral and central auditory system, or a combination of effects on both the auditory system and the high-level cognitive functions necessary for speech understanding in the subjects with high levels of noise exposure. As such, a test for listening effort would be useful as part of a comprehensive test battery within the assessment of subjects exposed to noise.
Collapse
|
30
|
Abstract
OBJECTIVES Whispered speech offers a unique set of challenges to speech perception and word recognition. The goals of the present study were twofold: First, to determine how listeners recognize whispered speech. Second, to inform major theories of spoken word recognition by considering how recognition changes when major cues to phoneme identity are reduced or largely absent compared with normal voiced speech. DESIGN Using eye tracking in the Visual World Paradigm, we examined how listeners recognize whispered speech. After hearing a target word (normal or whispered), participants selected the corresponding image from a display of four-a target (e.g., money), a word that shares sounds with the target at the beginning (cohort competitor, e.g., mother), a word that shares sounds with the target at the end (rhyme competitor, e.g., honey), and a phonologically unrelated word (e.g., whistle). Eye movements to each object were monitored to measure (1) how fast listeners process whispered speech, and (2) how strongly they consider lexical competitors (cohorts and rhymes) as the speech signal unfolds. RESULTS Listeners were slower to recognize whispered words. Compared with normal speech, listeners displayed slower reaction times to click the target image, were slower to fixate the target, and fixated the target less overall. Further, we found clear evidence that the dynamics of lexical competition are altered during whispered speech recognition. Relative to normal speech, words that overlapped with the target at the beginning (cohorts) displayed slower, reduced, and delayed activation, whereas words that overlapped with the target at the end (rhymes) exhibited faster, more robust, and longer lasting activation. CONCLUSION When listeners are confronted with whispered speech, they engage in a "wait-and-see" approach. Listeners delay lexical access, and by the time they begin to consider what word they are hearing, the beginning of the word has largely come and gone, and activation for cohorts is reduced. However, delays in lexical access actually increase consideration of rhyme competitors; the delay pushes lexical activation to a point later in processing, and the recognition system puts more weight on the word-final overlap between the target and the rhyme.
Collapse
|
31
|
Colby S, McMurray B. Cognitive and Physiological Measures of Listening Effort During Degraded Speech Perception: Relating Dual-Task and Pupillometry Paradigms. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:3627-3652. [PMID: 34491779 PMCID: PMC8642090 DOI: 10.1044/2021_jslhr-20-00583] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Revised: 04/01/2021] [Accepted: 05/21/2021] [Indexed: 06/13/2023]
Abstract
Purpose Listening effort is quickly becoming an important metric for assessing speech perception in less-than-ideal situations. However, the relationship between the construct of listening effort and the measures used to assess it remains unclear. We compared two measures of listening effort: a cognitive dual task and a physiological pupillometry task. We sought to investigate the relationship between these measures of effort and whether engaging effort impacts speech accuracy. Method In Experiment 1, 30 participants completed a dual task and a pupillometry task that were carefully matched in stimuli and design. The dual task consisted of a spoken word recognition task and a visual match-to-sample task. In the pupillometry task, pupil size was monitored while participants completed a spoken word recognition task. Both tasks presented words at three levels of listening difficulty (unmodified, eight-channel vocoding, and four-channel vocoding) and provided response feedback on every trial. We refined the pupillometry task in Experiment 2 (n = 31); crucially, participants no longer received response feedback. Finally, we ran a new group of subjects on both tasks in Experiment 3 (n = 30). Results In Experiment 1, accuracy in the visual task decreased with increased signal degradation in the dual task, but pupil size was sensitive to accuracy and not vocoding condition. After removing feedback in Experiment 2, changes in pupil size were predicted by listening condition, suggesting the task was now sensitive to engaged effort. Both tasks were sensitive to listening difficulty in Experiment 3, but there was no relationship between the tasks and neither task predicted speech accuracy. Conclusions Consistent with previous work, we found little evidence for a relationship between different measures of listening effort. We also found no evidence that effort predicts speech accuracy, suggesting that engaging more effort does not lead to improved speech recognition. Cognitive and physiological measures of listening effort are likely sensitive to different aspects of the construct of listening effort. Supplemental Material https://doi.org/10.23641/asha.16455900.
Collapse
Affiliation(s)
- Sarah Colby
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City
| | - Bob McMurray
- Department of Psychological and Brain Sciences, The University of Iowa, Iowa City
| |
Collapse
|
32
|
Icht M, Zukerman G, Ben-Itzchak E, Ben-David BM. Keep it simple: Identification of basic versus complex emotions in spoken language in individuals with autism spectrum disorder without intellectual disability: A meta-analysis study. Autism Res 2021; 14:1948-1964. [PMID: 34101373 DOI: 10.1002/aur.2551] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Revised: 03/14/2021] [Accepted: 05/04/2021] [Indexed: 12/26/2022]
Abstract
Daily functioning involves identifying emotions in spoken language, a fundamental aspect of social interactions. To date, there is inconsistent evidence in the literature on whether individuals with autism spectrum disorder without intellectual disability (ASD-without-ID) experience difficulties in identification of spoken emotions. We conducted a meta-analysis (literature search following the PRISMA guidelines), with 26 data sets (taken from 23 peer-reviewed journal articles) comparing individuals with ASD-without-ID (N = 614) and typically-developed (TD) controls (N = 640), from nine countries and in seven languages (published until February 2020). In our analyses there was no sufficient evidence to suggest that individuals with HF-ASD differ from matched controls in the identification of simple prosodic emotions (e.g., sadness, happiness). However, individuals with ASD-without-ID were found to perform significantly worse than controls in identification of complex prosodic emotions (e.g., envy and boredom). The level of the semantic content of the stimuli presented (e.g., sentences vs. strings of digits) was not found to have an impact on the results. In conclusion, the difference in findings between simple and complex emotions calls for a new-look on emotion processing in ASD-without-ID. Intervention programs may rely on the intact abilities of individuals with ASD-without-ID to process simple emotions and target improved performance with complex emotions. LAY SUMMARY: Individuals with autism spectrum disorder without intellectual disability (ASD-without-ID) do not differ from matched controls in the identification of simple prosodic emotions (e.g., sadness, happiness). However, they were found to perform significantly worse than controls in the identification of complex prosodic emotions (e.g., envy, boredom). This was found in a meta-analysis of 26 data sets with 1254 participants from nine countries and in seven languages. Intervention programs may rely on the intact abilities of individuals with ASD-without-ID to process simple emotions.
Collapse
Affiliation(s)
- Michal Icht
- Department of Communication Disorders, Ariel University, Ariel, Israel
| | - Gil Zukerman
- Department of Communication Disorders, Ariel University, Ariel, Israel
| | - Esther Ben-Itzchak
- Department of Communication Disorders, Ariel University, Ariel, Israel.,The Bruckner Center for Research in Autism, Department of Communication Disorders, Ariel University, Ariel, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC) Herzliya, Herzliya, Israel.,Department of Speech-Language Pathology, and Rehabilitation Sciences Institute (RSI), University of Toronto, Toronto, Ontario, Canada.,Toronto Rehabilitation Institute, University Health Networks (UHN), Toronto, Ontario, Canada
| |
Collapse
|
33
|
Tucker BV, Ford C, Hedges S. Speech aging: Production and perception. WILEY INTERDISCIPLINARY REVIEWS. COGNITIVE SCIENCE 2021; 12:e1557. [PMID: 33651922 DOI: 10.1002/wcs.1557] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2020] [Revised: 12/18/2020] [Accepted: 02/05/2021] [Indexed: 11/06/2022]
Abstract
In this overview we describe literature on how speech production and speech perception change in healthy or normal aging across the adult lifespan. In the production section we review acoustic characteristics that have been investigated as potentially distinguishing younger and older adults. In the speech perception section studies concerning speaker age estimation and those investigating older listeners' perception are addressed. Our discussion focuses on major themes and other fruitful areas for future research. This article is categorized under: Linguistics > Language in Mind and Brain Linguistics > Linguistic Theory Psychology > Development and Aging.
Collapse
Affiliation(s)
- Benjamin V Tucker
- Department of Linguistics, University of Alberta, Edmonton, Alberta, Canada
| | - Catherine Ford
- Department of Linguistics, University of Alberta, Edmonton, Alberta, Canada
| | - Stephanie Hedges
- Department of Linguistics, University of Alberta, Edmonton, Alberta, Canada
| |
Collapse
|
34
|
Kim S, Schwalje AT, Liu AS, Gander PE, McMurray B, Griffiths TD, Choi I. Pre- and post-target cortical processes predict speech-in-noise performance. Neuroimage 2021; 228:117699. [PMID: 33387631 PMCID: PMC8291856 DOI: 10.1016/j.neuroimage.2020.117699] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2020] [Revised: 11/06/2020] [Accepted: 12/23/2020] [Indexed: 12/19/2022] Open
Abstract
Understanding speech in noise (SiN) is a complex task that recruits multiple cortical subsystems. There is a variance in individuals' ability to understand SiN that cannot be explained by simple hearing profiles, which suggests that central factors may underlie the variance in SiN ability. Here, we elucidated a few cortical functions involved during a SiN task and their contributions to individual variance using both within- and across-subject approaches. Through our within-subject analysis of source-localized electroencephalography, we investigated how acoustic signal-to-noise ratio (SNR) alters cortical evoked responses to a target word across the speech recognition areas, finding stronger responses in left supramarginal gyrus (SMG, BA40 the dorsal lexicon area) with quieter noise. Through an individual differences approach, we found that listeners show different neural sensitivity to the background noise and target speech, reflected in the amplitude ratio of earlier auditory-cortical responses to speech and noise, named as an internal SNR. Listeners with better internal SNR showed better SiN performance. Further, we found that the post-speech time SMG activity explains a further amount of variance in SiN performance that is not accounted for by internal SNR. This result demonstrates that at least two cortical processes contribute to SiN performance independently: pre-target time processing to attenuate neural representation of background noise and post-target time processing to extract information from speech sounds.
Collapse
Affiliation(s)
- Subong Kim
- Department of Speech, Language, and Hearing Sciences, Purdue University, West Lafayette, IN 47907, USA
| | - Adam T Schwalje
- Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA
| | - Andrew S Liu
- Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA
| | - Phillip E Gander
- Department of Neurosurgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA
| | - Bob McMurray
- Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA; Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA 52242, USA; Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA 52242, USA
| | - Timothy D Griffiths
- Biosciences Institute, Newcastle University, Newcastle upon Tyne NE1 7RU, UK
| | - Inyong Choi
- Department of Otolaryngology - Head and Neck Surgery, University of Iowa Hospitals and Clinics, Iowa City, IA 52242, USA; Department of Communication Sciences and Disorders, University of Iowa, Iowa City, IA 52242, USA.
| |
Collapse
|
35
|
Harel-Arbeli T, Wingfield A, Palgi Y, Ben-David BM. Age-Related Differences in the Online Processing of Spoken Semantic Context and the Effect of Semantic Competition: Evidence From Eye Gaze. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:315-327. [PMID: 33561353 DOI: 10.1044/2020_jslhr-20-00142] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Purpose The study examined age-related differences in the use of semantic context and in the effect of semantic competition in spoken sentence processing. We used offline (response latency) and online (eye gaze) measures, using the "visual world" eye-tracking paradigm. Method Thirty younger and 30 older adults heard sentences related to one of four images presented on a computer monitor. They were asked to touch the image corresponding to the final word of the sentence (target word). Three conditions were used: a nonpredictive sentence, a predictive sentence suggesting one of the four images on the screen (semantic context), and a predictive sentence suggesting two possible images (semantic competition). Results Online eye gaze data showed no age-related differences with nonpredictive sentences, but revealed slowed processing for older adults when context was presented. With the addition of semantic competition to context, older adults were slower to look at the target word after it had been heard. In contrast, offline latency analysis did not show age-related differences in the effects of context and competition. As expected, older adults were generally slower to touch the image than younger adults. Conclusions Traditional offline measures were not able to reveal the complex effect of aging on spoken semantic context processing. Online eye gaze measures suggest that older adults were slower than younger adults to predict an indicated object based on semantic context. Semantic competition affected online processing for older adults more than for younger adults, with no accompanying age-related differences in latency. This supports an early age-related inhibition deficit, interfering with processing, and not necessarily with response execution.
Collapse
Affiliation(s)
- Tami Harel-Arbeli
- Department of Gerontology, University of Haifa, Israel
- Baruch Ivcher School of Psychology, Interdisciplinary Center Herzliya, Israel
| | - Arthur Wingfield
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA
| | - Yuval Palgi
- Department of Gerontology, University of Haifa, Israel
| | - Boaz M Ben-David
- Baruch Ivcher School of Psychology, Interdisciplinary Center Herzliya, Israel
- Department of Speech-Language Pathology, University of Toronto, Ontario, Canada
- Toronto Rehabilitation Institute, University Health Networks, Ontario, Canada
| |
Collapse
|
36
|
Hendrickson K, Oleson J, Walker E. School-Age Children Adapt the Dynamics of Lexical Competition in Suboptimal Listening Conditions. Child Dev 2021; 92:638-649. [PMID: 33476043 DOI: 10.1111/cdev.13530] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Abstract
Although the ability to understand speech in adverse listening conditions is paramount for effective communication across the life span, little is understood about how this critical processing skill develops. This study asks how the dynamics of spoken word recognition (i.e., lexical access and competition) change during soft speech in 8- to 11-year-olds (n = 26). Lexical competition and access for speech at lower intensity levels was measured using eye-tracking and the visual world paradigm. Overall the results suggest that soft speech influences the magnitude and timing of lexical access and competition. These results suggest that lexical competition is a cognitive process that can be adapted in the school-age years to help cope with increased uncertainty due to alterations in the speech signal.
Collapse
|
37
|
Berenbaum R, Tziraki C, Baum R, Rosen A, Reback T, Abikhzer J, Naparstek D, Ben-David BM. Focusing on Emotional and Social Intelligence Stimulation of People With Dementia by Playing a Serious Game—Proof of Concept Study. FRONTIERS IN COMPUTER SCIENCE 2020. [DOI: 10.3389/fcomp.2020.536880] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/09/2023] Open
Abstract
Context: Dementia is one of the top five chronic diseases, which has an overwhelming impact on patients' life quality, family, and caregivers. Currently, research relating to people with dementia (PwD) focuses on the deterioration of cognitive abilities. A more innovative approach, and the one taken by this paper, is focusing on methods to maintain and improve functionality, communication and quality of life of PwD by building on remaining capacities in the yet unexplored domain of emotional and social intelligence (ESI). The use of serious games for PwD (SG4D) aimed at building social and emotional capacity is a budding field of research.Objectives: Proof of concept that the, low cost, easy to deploy SG4D, called “My Brain Works” (MBW), co-designed with PwD, enhances ESI, based on the Bar-On ESI model.Methods: 27 PwD, clients at MELABEV dementia day center, participated in a mixed methods 12 weeks pilot, proof of concept study using a tablet SG4D co-designed with PwD. Quantitative performance data was collected automatically by the tablet during game sessions. In this paper we focus on the analysis of the qualitative and quantitative data related to ESI, observed by 10 different researchers, during each game session.Results: Quantitative data revealed: both the PwD with high and low MoCA scores had similar average ESI scores. Qualitative analysis revealed that the PwD demonstrated 9 sub-components of the Bar-On ESI Model.Conclusion: While there is no drug to stop cognitive decline associated with dementia, interventions related to ESI, on the other hand, may improve functioning and quality of life. Despite declines in cognitive abilities, our study shows that a tablet based SG4D can stimulate their ESI and evoke responses in self-awareness, empathy, social and communication capacities. Using SG4D to exercise and maintain social skills is an area that may be promising in the future and may help counter the negative effects of social isolation and loneliness. Such games, while not focusing on cognitive improvement, may also impact on cognitive functioning and help bridge the gap between caregiver and PwD. More research is needed with larger sample sizes.
Collapse
|
38
|
Xu J, Ke Y, Liu S, Song X, Xu C, Zhou G, Ming D. Task-irrelevant Auditory Event-related Potentials as Mental Workload Indicators: A Between-task Comparison Study . ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2020:3216-3219. [PMID: 33018689 DOI: 10.1109/embc44109.2020.9175957] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Real-time monitoring of mental workload (MWL) is a crucial step to build closed-loop adaptive aiding systems for human-machine systems. MWL estimators based on spontaneous electroencephalography (EEG) and event-related potentials (ERPs) have shown great potentials to achieve this goal. However, the previous studies show that the between-task robustness of these EEG/ERP-based MWL estimators is still an unsolved intractable question. This study attempts to examine the task-irrelevant auditory event-related potentials (tir-aERPs) as MWL indicators. A working memory task (verbal n-back) and a visuo-motor task (multi-attribute task battery, MATB), both with two difficulty levels (easy and hard), were used in the experiment, along with task-irrelevant auditory probes that did not need any response from the participants. EEG was recorded from ten participants when they were performing the tasks. The tir-aERPs elicited by the auditory probes were extracted and analyzed. The results show that the amplitudes of N1, early P3a (eP3a) and the late reorienting negativity (RON) significantly decreased with the increasing MWL in both n-back and MATB. Task type has no obvious influence on the amplitudes and topological layout of the MWL-sensitive tir-aERPs features. These results suggest that the tir-aERPs are potentially more constant MWL indicators across very different task types. Therefore, the tir-aERPs should be taken into consideration in future task-independent MWL monitoring studies.
Collapse
|
39
|
Ayasse ND, Wingfield A. The Two Sides of Linguistic Context: Eye-Tracking as a Measure of Semantic Competition in Spoken Word Recognition Among Younger and Older Adults. Front Hum Neurosci 2020; 14:132. [PMID: 32327987 PMCID: PMC7161414 DOI: 10.3389/fnhum.2020.00132] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2020] [Accepted: 03/20/2020] [Indexed: 12/17/2022] Open
Abstract
Studies of spoken word recognition have reliably shown that both younger and older adults' recognition of acoustically degraded words is facilitated by the presence of a linguistic context. Against this benefit, older adults' word recognition can be differentially hampered by interference from other words that could also fit the context. These prior studies have primarily used off-line response measures such as the signal-to-noise ratio needed for a target word to be correctly identified. Less clear is the locus of these effects; whether facilitation and interference have their influence primarily during response selection, or whether their effects begin to operate even before a sentence-final target word has been uttered. This question was addressed by tracking 20 younger and 20 older adults' eye fixations on a visually presented target word that corresponded to the final word of a contextually constraining or neutral sentence, accompanied by a second word on the computer screen that in some cases could also fit the sentence context. Growth curve analysis of the time-course of eye-gaze on a target word showed facilitation and inhibition effects begin to appear even as a spoken sentence is unfolding in time. Consistent with an age-related inhibition deficit, older adults' word recognition was slowed by the presence of a semantic competitor to a degree not observed for younger adults, with this effect operating early in the recognition process.
Collapse
Affiliation(s)
- Nicolai D Ayasse
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA, United States
| | - Arthur Wingfield
- Volen National Center for Complex Systems, Brandeis University, Waltham, MA, United States
| |
Collapse
|
40
|
Hendrickson K, Spinelli J, Walker E. Cognitive processes underlying spoken word recognition during soft speech. Cognition 2020; 198:104196. [PMID: 32004934 DOI: 10.1016/j.cognition.2020.104196] [Citation(s) in RCA: 16] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Revised: 01/06/2020] [Accepted: 01/18/2020] [Indexed: 11/25/2022]
Abstract
In two eye-tracking experiments using the Visual World Paradigm, we examined how listeners recognize words when faced with speech at lower intensities (40, 50, and 65 dBA). After hearing the target word, participants (n = 32) clicked the corresponding picture from a display of four images - a target (e.g., money), a cohort competitor (e.g., mother), a rhyme competitor (e.g., honey) and an unrelated item (e.g., whistle) - while their eye-movements were tracked. For slightly soft speech (50 dBA), listeners demonstrated an increase in cohort activation, whereas for rhyme competitors, activation started later and was sustained longer in processing. For very soft speech (40 dBA), listeners waited until later in processing to activate potential words, as illustrated by a decrease in activation for cohorts, and an increase in activation for rhymes. Further, the extent to which words were considered depended on word length (mono- vs. bi-syllabic words), and speech-extrinsic factors such as the surrounding listening environment. These results advance current theories of spoken word recognition by considering a range of speech levels more typical of everyday listening environments. From an applied perspective, these results motivate models of how individuals who are hard of hearing approach the task of recognizing spoken words.
Collapse
Affiliation(s)
- Kristi Hendrickson
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America; Department of Psychological & Brain Sciences, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| | - Jessica Spinelli
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| | - Elizabeth Walker
- Department of Communication Sciences & Disorders, University of Iowa, 250 Hawkins Drive, 52242 Iowa City, IA, United States of America.
| |
Collapse
|
41
|
Lewis GA, Bidelman GM. Autonomic Nervous System Correlates of Speech Categorization Revealed Through Pupillometry. Front Neurosci 2020; 13:1418. [PMID: 31998068 PMCID: PMC6967406 DOI: 10.3389/fnins.2019.01418] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2019] [Accepted: 12/16/2019] [Indexed: 02/06/2023] Open
Abstract
Human perception requires the many-to-one mapping between continuous sensory elements and discrete categorical representations. This grouping operation underlies the phenomenon of categorical perception (CP)-the experience of perceiving discrete categories rather than gradual variations in signal input. Speech perception requires CP because acoustic cues do not share constant relations with perceptual-phonetic representations. Beyond facilitating perception of unmasked speech, we reasoned CP might also aid the extraction of target speech percepts from interfering sound sources (i.e., noise) by generating additional perceptual constancy and reducing listening effort. Specifically, we investigated how noise interference impacts cognitive load and perceptual identification of unambiguous (i.e., categorical) vs. ambiguous stimuli. Listeners classified a speech vowel continuum (/u/-/a/) at various signal-to-noise ratios (SNRs [unmasked, 0 and -5 dB]). Continuous recordings of pupil dilation measured processing effort, with larger, later dilations reflecting increased listening demand. Critical comparisons were between time-locked changes in eye data in response to unambiguous (i.e., continuum endpoints) tokens vs. ambiguous tokens (i.e., continuum midpoint). Unmasked speech elicited faster responses and sharper psychometric functions, which steadily declined in noise. Noise increased pupil dilation across stimulus conditions, but not straightforwardly. Noise-masked speech modulated peak pupil size (i.e., [0 and -5 dB] > unmasked). In contrast, peak dilation latency varied with both token and SNR. Interestingly, categorical tokens elicited earlier pupil dilation relative to ambiguous tokens. Our pupillary data suggest CP reconstructs auditory percepts under challenging listening conditions through interactions between stimulus salience and listeners' internalized effort and/or arousal.
Collapse
Affiliation(s)
- Gwyneth A Lewis
- Institute for Intelligent Systems, The University of Memphis, Memphis, TN, United States.,School of Communication Sciences and Disorders, The University of Memphis, Memphis, TN, United States
| | - Gavin M Bidelman
- Institute for Intelligent Systems, The University of Memphis, Memphis, TN, United States.,School of Communication Sciences and Disorders, The University of Memphis, Memphis, TN, United States.,Department of Anatomy and Neurobiology, University of Tennessee Health Sciences Center, Memphis, TN, United States
| |
Collapse
|
42
|
Abstract
Recent applications of eye tracking for diagnosis, prognosis and follow-up of therapy in age-related neurological or psychological deficits have been reviewed. The review is focused on active aging, neurodegeneration and cognitive impairments. The potential impacts and current limitations of using characterizing features of eye movements and pupillary responses (oculometrics) as objective biomarkers in the context of aging are discussed. A closer look into the findings, especially with respect to cognitive impairments, suggests that eye tracking is an invaluable technique to study hidden aspects of aging that have not been revealed using any other noninvasive tool. Future research should involve a wider variety of oculometrics, in addition to saccadic metrics and pupillary responses, including nonlinear and combinatorial features as well as blink- and fixation-related metrics to develop biomarkers to trace age-related irregularities associated with cognitive and neural deficits.
Collapse
Affiliation(s)
- Ramtin Z Marandi
- Department of Health Science & Technology, Aalborg University, Aalborg E 9220, Denmark
| | - Parisa Gazerani
- Department of Health Science & Technology, Aalborg University, Aalborg E 9220, Denmark
| |
Collapse
|
43
|
Oron Y, Levy O, Avivi-Reich M, Goldfarb A, Handzel O, Shakuf V, Ben-David BM. Tinnitus affects the relative roles of semantics and prosody in the perception of emotions in spoken language. Int J Audiol 2019; 59:195-207. [DOI: 10.1080/14992027.2019.1677952] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Affiliation(s)
- Yahav Oron
- Department of Otolaryngology, Head, Neck and Maxillofacial Surgery, Tel-Aviv Sourasky Medical Center, Sackler School of Medicine, Tel Aviv University, Tel-Aviv, Israel
| | - Oren Levy
- Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC), Herzliya, Israel
| | - Meital Avivi-Reich
- Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC), Herzliya, Israel
- Communication Arts, Sciences and Disorders, Brooklyn College, City University of New York, New York, NY, USA
| | - Abraham Goldfarb
- Department of Otolaryngology, Head and Neck Surgery, The Edith Wolfson Medical Center, Sackler School of Medicine, Tel Aviv University, Tel-Aviv, Israel
| | - Ophir Handzel
- Department of Otolaryngology, Head, Neck and Maxillofacial Surgery, Tel-Aviv Sourasky Medical Center, Sackler School of Medicine, Tel Aviv University, Tel-Aviv, Israel
| | - Vered Shakuf
- Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC), Herzliya, Israel
| | - Boaz M. Ben-David
- Baruch Ivcher School of Psychology, Interdisciplinary Center (IDC), Herzliya, Israel
- Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada
- Toronto Rehabilitation Institute, University Health Networks (UHN), Toronto, ON, Canada
| |
Collapse
|
44
|
Nitsan G, Wingfield A, Lavie L, Ben-David BM. Differences in Working Memory Capacity Affect Online Spoken Word Recognition: Evidence From Eye Movements. Trends Hear 2019; 23:2331216519839624. [PMID: 31010398 PMCID: PMC6480998 DOI: 10.1177/2331216519839624] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022] Open
Abstract
Individual differences in working memory capacity have been gaining recognition as playing an important role in speech comprehension, especially in noisy environments. Using the visual world eye-tracking paradigm, a recent study by Hadar and coworkers found that online spoken word recognition was slowed when listeners were required to retain in memory a list of four spoken digits (high load) compared with only one (low load). In the current study, we recognized that the influence of a digit preload might be greater for individuals who have a more limited memory span. We compared participants with higher and lower memory spans on the time course for spoken word recognition by testing eye-fixations on a named object, relative to fixations on an object whose name shared phonology with the named object. Results show that when a low load was imposed, differences in memory span had no effect on the time course of preferential fixations. However, with a high load, listeners with lower span were delayed by ∼550 ms in discriminating target from sound-sharing competitors, relative to higher span listeners. This follows an assumption that the interference effect of a memory preload is not a fixed value, but rather, its effect is greater for individuals with a smaller memory span. Interestingly, span differences affected the timeline for spoken word recognition in noise, but not offline accuracy. This highlights the significance of using eye-tracking as a measure for online speech processing. Results further emphasize the importance of considering differences in cognitive capacity, even when testing normal hearing young adults.
Collapse
Affiliation(s)
- Gal Nitsan
- 1 Department of Communication Sciences and Disorders, University of Haifa, Israel.,2 Baruch Ivcher School of Psychology, Interdisciplinary Center Herzliya, Israel
| | - Arthur Wingfield
- 3 Volen National Center for Complex Systems, Brandeis University, Waltham, MA, USA
| | - Limor Lavie
- 1 Department of Communication Sciences and Disorders, University of Haifa, Israel
| | - Boaz M Ben-David
- 2 Baruch Ivcher School of Psychology, Interdisciplinary Center Herzliya, Israel.,4 Department of Speech-Language Pathology, University of Toronto, ON, Canada.,5 Toronto Rehabilitation Institute, University Health Networks, Toronto, ON, Canada
| |
Collapse
|
45
|
Mannheim I, Schwartz E, Xi W, Buttigieg SC, McDonnell-Naughton M, Wouters EJM, van Zaalen Y. Inclusion of Older Adults in the Research and Design of Digital Technology. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2019; 16:ijerph16193718. [PMID: 31581632 PMCID: PMC6801827 DOI: 10.3390/ijerph16193718] [Citation(s) in RCA: 84] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Received: 08/10/2019] [Revised: 09/24/2019] [Accepted: 09/30/2019] [Indexed: 01/09/2023]
Abstract
Digital technology holds a promise to improve older adults' well-being and promote ageing in place. However, there seems to be a discrepancy between digital technologies that are developed and what older adults actually want and need. Ageing is stereotypically framed as a problem needed to be fixed, and older adults are considered to be frail and incompetent. Not surprisingly, many of the technologies developed for the use of older adults focus on care. The exclusion of older adults from the research and design of digital technology is often based on such negative stereotypes. In this opinion article, we argue that the inclusion rather than exclusion of older adults in the design process and research of digital technology is essential if technology is to fulfill the promise of improving well-being. We emphasize why this is important while also providing guidelines, evidence from the literature, and examples on how to do so. We unequivocally state that designers and researchers should make every effort to ensure the involvement of older adults in the design process and research of digital technology. Based on this paper, we suggest that ageism in the design process of digital technology might play a role as a possible barrier of adopting technology.
Collapse
Affiliation(s)
- Ittay Mannheim
- School of Allied Health Professions, Fontys University of Applied Science, Eindhoven 5631 BN, The Netherlands.
- Tranzo, School of Social and Behavioral Sciences, Tilburg University, Tilburg 5037 DB, The Netherlands.
| | - Ella Schwartz
- Louis and Gabi Weisfeld School of Social Work, Bar Ilan University, Ramat Gan 5290002, Israel.
- Department of Public Health, Faculty of Health Sciences, Ben-Gurion University of the Negev, Be'er Sheva 8410501, Israel.
| | - Wanyu Xi
- Louis and Gabi Weisfeld School of Social Work, Bar Ilan University, Ramat Gan 5290002, Israel.
| | - Sandra C Buttigieg
- Health Services Management, Faculty of Health Sciences, University of Malta, Msida MSD 2090, Malta.
- Health Services Management Centre, School of Social Policy, College of Social Sciences, University of Birmingham, Edgbaston, Birmingham B15 2TT, UK.
| | - Mary McDonnell-Naughton
- Department of Nursing and Health Care Athlone Institute of Technology, Athlone N37 HD68, Ireland.
| | - Eveline J M Wouters
- School of Allied Health Professions, Fontys University of Applied Science, Eindhoven 5631 BN, The Netherlands.
- Tranzo, School of Social and Behavioral Sciences, Tilburg University, Tilburg 5037 DB, The Netherlands.
| | - Yvonne van Zaalen
- School of Allied Health Professions, Fontys University of Applied Science, Eindhoven 5631 BN, The Netherlands.
| |
Collapse
|
46
|
Psychobiological Responses Reveal Audiovisual Noise Differentially Challenges Speech Recognition. Ear Hear 2019; 41:268-277. [PMID: 31283529 DOI: 10.1097/aud.0000000000000755] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES In noisy environments, listeners benefit from both hearing and seeing a talker, demonstrating audiovisual (AV) cues enhance speech-in-noise (SIN) recognition. Here, we examined the relative contribution of auditory and visual cues to SIN perception and the strategies used by listeners to decipher speech in noise interference(s). DESIGN Normal-hearing listeners (n = 22) performed an open-set speech recognition task while viewing audiovisual TIMIT sentences presented under different combinations of signal degradation including visual (AVn), audio (AnV), or multimodal (AnVn) noise. Acoustic and visual noises were matched in physical signal-to-noise ratio. Eyetracking monitored participants' gaze to different parts of a talker's face during SIN perception. RESULTS As expected, behavioral performance for clean sentence recognition was better for A-only and AV compared to V-only speech. Similarly, with noise in the auditory channel (AnV and AnVn speech), performance was aided by the addition of visual cues of the talker regardless of whether the visual channel contained noise, confirming a multimodal benefit to SIN recognition. The addition of visual noise (AVn) obscuring the talker's face had little effect on speech recognition by itself. Listeners' eye gaze fixations were biased toward the eyes (decreased at the mouth) whenever the auditory channel was compromised. Fixating on the eyes was negatively associated with SIN recognition performance. Eye gazes on the mouth versus eyes of the face also depended on the gender of the talker. CONCLUSIONS Collectively, results suggest listeners (1) depend heavily on the auditory over visual channel when seeing and hearing speech and (2) alter their visual strategy from viewing the mouth to viewing the eyes of a talker with signal degradations, which negatively affects speech perception.
Collapse
|
47
|
Fostick L. Card playing enhances speech perception among aging adults: comparison with aging musicians. Eur J Ageing 2019; 16:481-489. [PMID: 31798372 DOI: 10.1007/s10433-019-00512-2] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/09/2023] Open
Abstract
Speech perception and auditory processing have been shown to be enhanced among aging musicians as compared to non-musicians. In the present study, the aim was to test whether these functions are also enhanced among those who are engaged in a non-musical mentally challenging leisure activity (card playing). Three groups of 23 aging adults, aged 60-80 years, were recruited for the study: Musicians, Card players, and Controls. Participants were matched for age, gender, Wechsler Adult Intelligence Scale-III Matrix Reasoning, and Digit Span scores. Their performance was measured using auditory spectral and spatial temporal order judgment tests, and four tasks of speech perception in conditions of: no background noise, background noise of speech frequencies, background noise of white noise, and 60% compressed speech. Musicians were better in auditory and speech perception than the other two groups. Card players were similar to Controls in auditory perception tasks, but were better in the speech perception tasks. Non-musician aging adults may be able to improve their speech perception ability by engaging in leisure activity requiring cognitive effort.
Collapse
Affiliation(s)
- Leah Fostick
- Department of Communication Disorders, Ariel University, Ariel, Israel
| |
Collapse
|
48
|
Ronen M, Lifshitz-Ben-Basat A, Taitelbaum-Swead R, Fostick L. Auditory temporal processing, reading, and phonological awareness among aging adults. Acta Psychol (Amst) 2018; 190:1-10. [PMID: 29986206 DOI: 10.1016/j.actpsy.2018.06.010] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2018] [Revised: 06/25/2018] [Accepted: 06/25/2018] [Indexed: 11/18/2022] Open
Abstract
Auditory temporal processing (ATP) has been related in the literature to both speech perception as well as reading and phonological awareness. In aging adults, it is known to be related to difficulties in speech perception. In the present study, we aimed to test whether an age-related deficit in ATP would also be accompanied by poor reading and phonological awareness. Thirty-eight aging adults were compared to 55 readers with dyslexia and 42 young normal readers on temporal order judgment (TOJ), speech perception, reading, and phonological awareness tests. Aging adults had longer TOJ thresholds than young normal readers, but shorter than readers with dyslexia; however, they had lower speech perception accuracy than both groups. Phonological awareness of the aging adults was better than readers with dyslexia, but poorer than young normal readers, although their reading accuracy was similar to that of the young controls. This is the first report on poor phonological awareness among aging adults. Suprisingly, it was not accompanied by difficulties in reading ability, and might instead be related to aging adults' difficulties in speech perception. This newly discovered relationship between ATP and phonological awareness among aging adults appears to extend the existing understanding of this relationship, and suggests it should be explored in other groups with ATP deficits.
Collapse
Affiliation(s)
- Michal Ronen
- Department of Psychology, Ariel University, Israel
| | | | | | - Leah Fostick
- Department of Communication Disorders, Ariel University, Israel.
| |
Collapse
|
49
|
Van Engen KJ, McLaughlin DJ. Eyes and ears: Using eye tracking and pupillometry to understand challenges to speech recognition. Hear Res 2018; 369:56-66. [PMID: 29801981 DOI: 10.1016/j.heares.2018.04.013] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/03/2017] [Revised: 04/12/2018] [Accepted: 04/25/2018] [Indexed: 11/16/2022]
Abstract
Although human speech recognition is often experienced as relatively effortless, a number of common challenges can render the task more difficult. Such challenges may originate in talkers (e.g., unfamiliar accents, varying speech styles), the environment (e.g. noise), or in listeners themselves (e.g., hearing loss, aging, different native language backgrounds). Each of these challenges can reduce the intelligibility of spoken language, but even when intelligibility remains high, they can place greater processing demands on listeners. Noisy conditions, for example, can lead to poorer recall for speech, even when it has been correctly understood. Speech intelligibility measures, memory tasks, and subjective reports of listener difficulty all provide critical information about the effects of such challenges on speech recognition. Eye tracking and pupillometry complement these methods by providing objective physiological measures of online cognitive processing during listening. Eye tracking records the moment-to-moment direction of listeners' visual attention, which is closely time-locked to unfolding speech signals, and pupillometry measures the moment-to-moment size of listeners' pupils, which dilate in response to increased cognitive load. In this paper, we review the uses of these two methods for studying challenges to speech recognition.
Collapse
|
50
|
Keller CH, Kaylegian K, Wehr M. Gap encoding by parvalbumin-expressing interneurons in auditory cortex. J Neurophysiol 2018; 120:105-114. [PMID: 29589814 DOI: 10.1152/jn.00911.2017] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
Synaptic inhibition shapes the temporal processing of sounds in auditory cortex, but the contribution of specific inhibitory cell types to temporal processing remains unclear. We recorded from parvalbumin-expressing (PV+) interneurons in auditory cortex to determine how they encode gaps in noise, a model of temporal processing more generally. We found that PV+ cells had stronger and more prevalent on-responses, off-responses, and postresponse suppression compared with presumed pyramidal cells. We summarize this pattern of differences as "deeper modulation" of gap responses in PV+ cells. Response latencies were also markedly faster for PV+ cells. We found a similar pattern of deeper modulation and faster latencies for responses to white noise bursts, suggesting that these are general properties of on- and off-responses in PV+ cells rather than specific features of gap encoding. These findings are consistent with a role for PV+ cells in providing dynamic gain control by pooling local activity. NEW & NOTEWORTHY We found that parvalbumin-expressing (PV+) interneurons in auditory cortex showed more deeply modulated responses to both gaps in noise and bursts of noise, suggesting that they are optimized for the rapid detection of stimulus transients.
Collapse
Affiliation(s)
| | | | - Michael Wehr
- Institute of Neuroscience, University of Oregon , Eugene, Oregon
| |
Collapse
|