1
|
Slomianka V, Dau T, Ahrens A. Acoustic scene complexity affects motion behavior during speech perception in audio-visual multi-talker virtual environments. Sci Rep 2024; 14:19028. [PMID: 39152193 PMCID: PMC11329770 DOI: 10.1038/s41598-024-70026-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2024] [Accepted: 08/09/2024] [Indexed: 08/19/2024] Open
Abstract
In real-world listening situations, individuals typically utilize head and eye movements to receive and enhance sensory information while exploring acoustic scenes. However, the specific patterns of such movements have not yet been fully characterized. Here, we studied how movement behavior is influenced by scene complexity, varied in terms of reverberation and the number of concurrent talkers. Thirteen normal-hearing participants engaged in a speech comprehension and localization task, requiring them to indicate the spatial location of a spoken story in the presence of other stories in virtual audio-visual scenes. We observed delayed initial head movements when more simultaneous talkers were present in the scene. Both reverberation and a higher number of talkers extended the search period, increased the number of fixated source locations, and resulted in more gaze jumps. The period preceding the participants' responses was prolonged when more concurrent talkers were present, and listeners continued to move their eyes in the proximity of the target talker. In scenes with more reverberation, the final head position when making the decision was farther away from the target. These findings demonstrate that the complexity of the acoustic scene influences listener behavior during speech comprehension and localization in audio-visual scenes.
Collapse
Affiliation(s)
- Valeska Slomianka
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, 2800, Kgs. Lyngby, Denmark.
| | - Torsten Dau
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, 2800, Kgs. Lyngby, Denmark
| | - Axel Ahrens
- Hearing Systems Section, Department of Health Technology, Technical University of Denmark, 2800, Kgs. Lyngby, Denmark
| |
Collapse
|
2
|
Miles K, Weisser A, Kallen RW, Varlet M, Richardson MJ, Buchholz JM. Behavioral dynamics of conversation, (mis)communication and coordination in noisy environments. Sci Rep 2023; 13:20271. [PMID: 37985887 PMCID: PMC10662155 DOI: 10.1038/s41598-023-47396-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Accepted: 11/13/2023] [Indexed: 11/22/2023] Open
Abstract
During conversations people coordinate simultaneous channels of verbal and nonverbal information to hear and be heard. But the presence of background noise levels such as those found in cafes and restaurants can be a barrier to conversational success. Here, we used speech and motion-tracking to reveal the reciprocal processes people use to communicate in noisy environments. Conversations between twenty-two pairs of typical-hearing adults were elicited under different conditions of background noise, while standing or sitting around a table. With the onset of background noise, pairs rapidly adjusted their interpersonal distance and speech level, with the degree of initial change dependent on noise level and talker configuration. Following this transient phase, pairs settled into a sustaining phase in which reciprocal speech and movement-based coordination processes synergistically maintained effective communication, again with the magnitude of stability of these coordination processes covarying with noise level and talker configuration. Finally, as communication breakdowns increased at high noise levels, pairs exhibited resetting behaviors to help restore communication-decreasing interpersonal distance and/or increasing speech levels in response to communication breakdowns. Approximately 78 dB SPL defined a threshold where behavioral processes were no longer sufficient for maintaining effective conversation and communication breakdowns rapidly increased.
Collapse
Affiliation(s)
- Kelly Miles
- ECHO Laboratory, MU Hearing, and Performance and Expertise Research Centre, Macquarie University, Sydney, Australia.
| | - Adam Weisser
- ECHO Laboratory, Macquarie University, Sydney, Australia
| | - Rachel W Kallen
- Performance and Expertise Research Centre, School of Psychological Sciences, Macquarie University, Sydney, Australia
| | - Manuel Varlet
- The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Sydney, Australia
| | - Michael J Richardson
- Performance and Expertise Research Centre, School of Psychological Sciences, Macquarie University, Sydney, Australia
| | - Joerg M Buchholz
- ECHO Laboratory, MU Hearing, and Performance and Expertise Research Centre, Macquarie University, Sydney, Australia
| |
Collapse
|
3
|
Tsai MJ. Dyadic Conversation between Mandarin-Chinese-Speaking Healthy Older Adults: From Analyses of Conversation Turns and Speaking Roles. Behav Sci (Basel) 2023; 13:bs13020134. [PMID: 36829363 PMCID: PMC9952709 DOI: 10.3390/bs13020134] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2022] [Revised: 01/27/2023] [Accepted: 02/03/2023] [Indexed: 02/09/2023] Open
Abstract
Older adults' daily conversations with other older adults enable them to connect to their surrounding communities and improve their friendships. However, typical aging processes and fluctuations in family caring might cause conversation changes. The purpose of this study was to explore the quantitative contributions of conversation turns (CTs) and speaking roles (SRs) in Mandarin-Chinese-speaking conversation dyads between mutually familiar healthy older adults (HOAs). A total of 20 HOAs aged 65 or over were recruited. Each dyad conversed for ten minutes once a week for five weeks, five sessions per dyad, for a total of 50 sessions. The frequency and percentages of the coded CTs and SRs contributed by each HOA were individually tallied and calculated. Quantitatively symmetrical contributions of CTs and SRs occurred in Mandarin-Chinese-speaking conversation dyads between mutually familiar HOAs. Although typical aging processes might change conversations, both Mandarin-Chinese-speaking HOAs serve as active interlocutors to each other in taking CTs and SRs to co-construct their conversation processes and content in their dyadic conversation. Sufficient knowledge of conversation co-constructions might lead them to have more supportive environments to connect to surrounding communities and improve their friendships.
Collapse
Affiliation(s)
- Meng-Ju Tsai
- Department of Speech-Language Pathology and Audiology, Chung Shan Medical University, Taichung City 402, Taiwan;
- Speech and Language Therapy Room, Chung Shan Medical University Hospital, Taichung City 402, Taiwan
| |
Collapse
|
4
|
Valzolgher C, Alzaher M, Gaveau V, Coudert A, Marx M, Truy E, Barone P, Farnè A, Pavani F. Capturing Visual Attention With Perturbed Auditory Spatial Cues. Trends Hear 2023; 27:23312165231182289. [PMID: 37611181 PMCID: PMC10467228 DOI: 10.1177/23312165231182289] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2022] [Revised: 05/25/2023] [Accepted: 05/29/2023] [Indexed: 08/25/2023] Open
Abstract
Lateralized sounds can orient visual attention, with benefits for audio-visual processing. Here, we asked to what extent perturbed auditory spatial cues-resulting from cochlear implants (CI) or unilateral hearing loss (uHL)-allow this automatic mechanism of information selection from the audio-visual environment. We used a classic paradigm from experimental psychology (capture of visual attention with sounds) to probe the integrity of audio-visual attentional orienting in 60 adults with hearing loss: bilateral CI users (N = 20), unilateral CI users (N = 20), and individuals with uHL (N = 20). For comparison, we also included a group of normal-hearing (NH, N = 20) participants, tested in binaural and monaural listening conditions (i.e., with one ear plugged). All participants also completed a sound localization task to assess spatial hearing skills. Comparable audio-visual orienting was observed in bilateral CI, uHL, and binaural NH participants. By contrast, audio-visual orienting was, on average, absent in unilateral CI users and reduced in NH listening with one ear plugged. Spatial hearing skills were better in bilateral CI, uHL, and binaural NH participants than in unilateral CI users and monaurally plugged NH listeners. In unilateral CI users, spatial hearing skills correlated with audio-visual-orienting abilities. These novel results show that audio-visual-attention orienting can be preserved in bilateral CI users and in uHL patients to a greater extent than unilateral CI users. This highlights the importance of assessing the impact of hearing loss beyond auditory difficulties alone: to capture to what extent it may enable or impede typical interactions with the multisensory environment.
Collapse
Affiliation(s)
- Chiara Valzolgher
- Center for Mind/Brain Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team, Lyon Neuroscience Research Center, Lyon, France
| | - Mariam Alzaher
- Centre de Recherche Cerveau & Cognition, Toulouse, France
- Hospices Civils, Toulouse, France
| | - Valérie Gaveau
- Integrative, Multisensory, Perception, Action and Cognition Team, Lyon Neuroscience Research Center, Lyon, France
| | | | - Mathieu Marx
- Centre de Recherche Cerveau & Cognition, Toulouse, France
- Hospices Civils, Toulouse, France
| | - Eric Truy
- Integrative, Multisensory, Perception, Action and Cognition Team, Lyon Neuroscience Research Center, Lyon, France
- Hospices Civils de Lyon, Lyon, France
| | - Pascal Barone
- Centre de Recherche Cerveau & Cognition, Toulouse, France
| | - Alessandro Farnè
- Center for Mind/Brain Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team, Lyon Neuroscience Research Center, Lyon, France
- Neuro-immersion, Lyon, France
| | - Francesco Pavani
- Center for Mind/Brain Sciences - CIMeC, University of Trento, Rovereto, Italy
- Integrative, Multisensory, Perception, Action and Cognition Team, Lyon Neuroscience Research Center, Lyon, France
- Centro Interuniversitario di Ricerca « Cognizione, Linguaggio e Sordità », Rovereto, Italy
| |
Collapse
|
5
|
Hadley LV, Culling JF. Timing of head turns to upcoming talkers in triadic conversation: Evidence for prediction of turn ends and interruptions. Front Psychol 2022; 13:1061582. [PMID: 36605274 PMCID: PMC9807761 DOI: 10.3389/fpsyg.2022.1061582] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2022] [Accepted: 11/24/2022] [Indexed: 12/24/2022] Open
Abstract
In conversation, people are able to listen to an utterance and respond within only a few hundred milliseconds. It takes substantially longer to prepare even a simple utterance, suggesting that interlocutors may make use of predictions about when the talker is about to end. But it is not only the upcoming talker that needs to anticipate the prior talker ending-listeners that are simply following the conversation could also benefit from predicting the turn end in order to shift attention appropriately with the turn switch. In this paper, we examined whether people predict upcoming turn ends when watching conversational turns switch between others by analysing natural conversations. These conversations were between triads of older adults in different levels and types of noise. The analysis focused on the observer during turn switches between the other two parties using head orientation (i.e. saccades from one talker to the next) to identify when their focus moved from one talker to the next. For non-overlapping utterances, observers started to turn to the upcoming talker before the prior talker had finished speaking in 17% of turn switches (going up to 26% when accounting for motor-planning time). For overlapping utterances, observers started to turn towards the interrupter before they interrupted in 18% of turn switches (going up to 33% when accounting for motor-planning time). The timing of head turns was more precise at lower than higher noise levels, and was not affected by noise type. These findings demonstrate that listeners in natural group conversation situations often exhibit head movements that anticipate the end of one conversational turn and the beginning of another. Furthermore, this work demonstrates the value of analysing head movement as a cue to social attention, which could be relevant for advancing communication technology such as hearing devices.
Collapse
Affiliation(s)
- Lauren V. Hadley
- Hearing Sciences – Scottish Section, School of Medicine, University of Nottingham, Glasgow, United Kingdom
| | - John F. Culling
- School of Psychology, Cardiff University, Cardiff, United Kingdom
| |
Collapse
|
6
|
Katiri R, Sivan N, Noone A, Farrell E, McLoughlin L, Lang B, O'Donnell B, Kieran SM. Outcomes from 7 years of a direct to audiology referral pathway. Ir J Med Sci 2022:10.1007/s11845-022-03145-0. [PMID: 36056228 PMCID: PMC9439272 DOI: 10.1007/s11845-022-03145-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2022] [Accepted: 08/28/2022] [Indexed: 11/06/2022]
Abstract
Background The direct to audiology via ENT referral pathway was designed to enhance patient access to audiology services. The pathway is recommended in the Otolaryngology Head & Neck Surgery: a model of care for Ireland report, published in 2019. Aims This study aimed to review the outcomes of all patients that attended our institution over the last seven years. Methods A retrospective review of the direct to audiology referral service from March 2014 to December 2021 was conducted. Outcomes were defined by the pathway each patient took following attendance at the audiology clinic. Patients were either (i) discharged, (ii) referred for hearing aid(s) or (iii) identified as candidates for further diagnostic assessments, including a follow-up at the ENT outpatient clinic. Results During the time frame, 419 patients were triaged to the pathway. The average wait time was 13 days. The average age was 53 years (range 16–96 years, SD = 6.1). Approximately 34% (n = 143) of all patients referred were discharged back to the GP by the audiologist, but 66% (n = 276) presented with ‘red flags’ and needed further investigation in the ENT clinic, with 30% (n = 73) ultimately requiring imaging studies. Over half (n = 254, 61%) were referred for hearing aids. Conclusion The direct to audiology initiative has proven effective at reducing waiting times for ENT patients solely in need of audiological intervention. Approximately one-third of these referrals to the ENT service can be assessed comprehensively in the audiology clinic, thereby reducing the demand for ENT clinics, enhancing service provision and expediting onwards referral for amplification.
Collapse
Affiliation(s)
- Roulla Katiri
- Department of Audiology, Mater Misericordiae University Hospital, Dublin, Ireland. .,Hearing Sciences, Mental Health and Clinical Neuroscience, School of Medicine, University of Nottingham, Nottingham, UK.
| | - Nina Sivan
- Department of Audiology, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Anthony Noone
- ENT Department, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Eric Farrell
- ENT Department, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Laura McLoughlin
- ENT Department, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Bronagh Lang
- ENT Department, Mater Misericordiae University Hospital, Dublin, Ireland
| | - Bronagh O'Donnell
- HSE Community Audiology Services, Grangegorman Primary Care Centre, Dublin, Ireland
| | - Stephen M Kieran
- ENT Department, Mater Misericordiae University Hospital, Dublin, Ireland
| |
Collapse
|
7
|
Hidalgo C, Mohamed I, Zielinski C, Schön D. The effect of speech degradation on the ability to track and predict turn structure in conversation. Cortex 2022; 151:105-115. [DOI: 10.1016/j.cortex.2022.01.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2021] [Revised: 11/15/2021] [Accepted: 01/20/2022] [Indexed: 11/03/2022]
|
8
|
Hadley LV, Ward JA. Synchrony as a measure of conversation difficulty: Movement coherence increases with background noise level and complexity in dyads and triads. PLoS One 2021; 16:e0258247. [PMID: 34610018 PMCID: PMC8491905 DOI: 10.1371/journal.pone.0258247] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2020] [Accepted: 09/22/2021] [Indexed: 11/18/2022] Open
Abstract
When people interact, they fall into synchrony. This synchrony has been demonstrated in a range of contexts, from walking or playing music together to holding a conversation, and has been linked to prosocial outcomes such as development of rapport and efficiency of cooperation. While the basis of synchrony remains unclear, several studies have found synchrony to increase when an interaction is made challenging, potentially providing a means of facilitating interaction. Here we focus on head movement during free conversation. As verbal information is obscured when conversing over background noise, we investigate whether synchrony is greater in high vs low levels of noise, as well as addressing the effect of background noise complexity. Participants held a series of conversations with unfamiliar interlocutors while seated in a lab, and the background noise level changed every 15-30s between 54, 60, 66, 72, and 78 dB. We report measures of head movement synchrony recorded via high-resolution motion tracking at the extreme noise levels (i.e., 54 vs 78 dB) in dyads (n = 15) and triads (n = 11). In both the dyads and the triads, we report increased movement coherence in high compared to low level speech-shaped noise. Furthermore, in triads we compare behaviour in speech-shaped noise vs multi-talker babble, and find greater movement coherence in the more complex babble condition. Key synchrony differences fall in the 0.2-0.5 Hz frequency bands, and are discussed in terms of their correspondence to talkers' average utterance durations. Additional synchrony differences occur at higher frequencies in the triads only (i.e., >5 Hz), which may relate to synchrony of backchannel cues (as multiple individuals were listening and responding to the same talker). Not only do these studies replicate prior work indicating interlocutors' increased reliance on behavioural synchrony as task difficulty increases, but they demonstrate these effects using multiple difficulty manipulations and across different sized interaction groups.
Collapse
Affiliation(s)
- Lauren V. Hadley
- Hearing Sciences—Scottish Section, School of Medicine, University of Nottingham, Glasgow, United Kingdom
- * E-mail:
| | - Jamie A. Ward
- Computing, Goldsmiths, University of London, London, United Kingdom
| |
Collapse
|