1
|
Bauer A, Kuder A, Schulder M, Schepens J. Phonetic differences between affirmative and feedback head nods in German Sign Language (DGS): A pose estimation study. PLoS One 2024; 19:e0304040. [PMID: 38814896 PMCID: PMC11139280 DOI: 10.1371/journal.pone.0304040] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2024] [Accepted: 05/04/2024] [Indexed: 06/01/2024] Open
Abstract
This study investigates head nods in natural dyadic German Sign Language (DGS) interaction, with the aim of finding whether head nods serving different functions vary in their phonetic characteristics. Earlier research on spoken and sign language interaction has revealed that head nods vary in the form of the movement. However, most claims about the phonetic properties of head nods have been based on manual annotation without reference to naturalistic text types and the head nods produced by the addressee have been largely ignored. There is a lack of detailed information about the phonetic properties of the addressee's head nods and their interaction with manual cues in DGS as well as in other sign languages, and the existence of a form-function relationship of head nods remains uncertain. We hypothesize that head nods functioning in the context of affirmation differ from those signaling feedback in their form and the co-occurrence with manual items. To test the hypothesis, we apply OpenPose, a computer vision toolkit, to extract head nod measurements from video recordings and examine head nods in terms of their duration, amplitude and velocity. We describe the basic phonetic properties of head nods in DGS and their interaction with manual items in naturalistic corpus data. Our results show that phonetic properties of affirmative nods differ from those of feedback nods. Feedback nods appear to be on average slower in production and smaller in amplitude than affirmation nods, and they are commonly produced without a co-occurring manual element. We attribute the variations in phonetic properties to the distinct roles these cues fulfill in turn-taking system. This research underlines the importance of non-manual cues in shaping the turn-taking system of sign languages, establishing the links between such research fields as sign language linguistics, conversational analysis, quantitative linguistics and computer vision.
Collapse
Affiliation(s)
- Anastasia Bauer
- Department of Linguistics, General Linguistics, University of Cologne, Cologne, Germany
| | - Anna Kuder
- Department of Linguistics, General Linguistics, University of Cologne, Cologne, Germany
| | - Marc Schulder
- Institute for German Sign Language and Communication of the Deaf, University of Hamburg, Hamburg, Germany
| | - Job Schepens
- Department of Linguistics, General Linguistics, University of Cologne, Cologne, Germany
| |
Collapse
|
2
|
Casartelli L, Maronati C, Cavallo A. From neural noise to co-adaptability: Rethinking the multifaceted architecture of motor variability. Phys Life Rev 2023; 47:245-263. [PMID: 37976727 DOI: 10.1016/j.plrev.2023.10.036] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 10/27/2023] [Indexed: 11/19/2023]
Abstract
In the last decade, the source and the functional meaning of motor variability have attracted considerable attention in behavioral and brain sciences. This construct classically combined different levels of description, variable internal robustness or coherence, and multifaceted operational meanings. We provide here a comprehensive review of the literature with the primary aim of building a precise lexicon that goes beyond the generic and monolithic use of motor variability. In the pars destruens of the work, we model three domains of motor variability related to peculiar computational elements that influence fluctuations in motor outputs. Each domain is in turn characterized by multiple sub-domains. We begin with the domains of noise and differentiation. However, the main contribution of our model concerns the domain of adaptability, which refers to variation within the same exact motor representation. In particular, we use the terms learning and (social)fitting to specify the portions of motor variability that depend on our propensity to learn and on our largely constitutive propensity to be influenced by external factors. A particular focus is on motor variability in the context of the sub-domain named co-adaptability. Further groundbreaking challenges arise in the modeling of motor variability. Therefore, in a separate pars construens, we attempt to characterize these challenges, addressing both theoretical and experimental aspects as well as potential clinical implications for neurorehabilitation. All in all, our work suggests that motor variability is neither simply detrimental nor beneficial, and that studying its fluctuations can provide meaningful insights for future research.
Collapse
Affiliation(s)
- Luca Casartelli
- Theoretical and Cognitive Neuroscience Unit, Scientific Institute IRCCS E. MEDEA, Italy
| | - Camilla Maronati
- Move'n'Brains Lab, Department of Psychology, Università degli Studi di Torino, Italy
| | - Andrea Cavallo
- Move'n'Brains Lab, Department of Psychology, Università degli Studi di Torino, Italy; C'MoN Unit, Fondazione Istituto Italiano di Tecnologia, Genova, Italy.
| |
Collapse
|
3
|
Raghavan R, Raviv L, Peeters D. What's your point? Insights from virtual reality on the relation between intention and action in the production of pointing gestures. Cognition 2023; 240:105581. [PMID: 37573692 DOI: 10.1016/j.cognition.2023.105581] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Revised: 07/03/2023] [Accepted: 07/26/2023] [Indexed: 08/15/2023]
Abstract
Human communication involves the process of translating intentions into communicative actions. But how exactly do our intentions surface in the visible communicative behavior we display? Here we focus on pointing gestures, a fundamental building block of everyday communication, and investigate whether and how different types of underlying intent modulate the kinematics of the pointing hand and the brain activity preceding the gestural movement. In a dynamic virtual reality environment, participants pointed at a referent to either share attention with their addressee, inform their addressee, or get their addressee to perform an action. Behaviorally, it was observed that these different underlying intentions modulated how long participants kept their arm and finger still, both prior to starting the movement and when keeping their pointing hand in apex position. In early planning stages, a neurophysiological distinction was observed between a gesture that is used to share attitudes and knowledge with another person versus a gesture that mainly uses that person as a means to perform an action. Together, these findings suggest that our intentions influence our actions from the earliest neurophysiological planning stages to the kinematic endpoint of the movement itself.
Collapse
Affiliation(s)
- Renuka Raghavan
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands; Radboud University, Donders Institute for Brain, Cognition, and Behavior, Nijmegen, The Netherlands
| | - Limor Raviv
- Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands; Centre for Social, Cognitive and Affective Neuroscience (cSCAN), University of Glasgow, United Kingdom
| | - David Peeters
- Tilburg University, Department of Communication and Cognition, TiCC, Tilburg, The Netherlands.
| |
Collapse
|
4
|
Quarmley M, Zelinsky G, Athar S, Yang Z, Drucker JH, Samaras D, Jarcho JM. Nonverbal behavioral patterns predict social rejection elicited aggression. Biol Psychol 2023; 183:108670. [PMID: 37652178 PMCID: PMC10591947 DOI: 10.1016/j.biopsycho.2023.108670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2023] [Revised: 08/02/2023] [Accepted: 08/28/2023] [Indexed: 09/02/2023]
Abstract
Aggression elicited by social rejection is costly, prevalent, and often lethal. Attempts to predict rejection-elicited aggression using trait-based data have had little success. This may be because in-the-moment aggression is a complex process influenced by current states of attention, arousal, and affect which are poorly predicted by trait-level characteristics. In a study of young adults (N = 89; 18-25 years), machine learning tested the extent to which nonverbal behavioral indices of attention (eye gaze), arousal (pupillary reactivity), and affect (facial expressions) during a novel social interaction paradigm predicted subsequent aggression towards rejecting and accepting peers. Eye gaze and pupillary reactivity predicted aggressive behavior; predictions were more successful than measures of trait-based aggression and harsh parenting. These preliminary results suggest that nonverbal behavior may elucidate underlying mechanisms of in-the-moment aggression.
Collapse
Affiliation(s)
- M Quarmley
- Department of Psychology, Temple University, Philadelphia, PA, United States
| | - G Zelinsky
- Department of Psychology, Stony Brook University, Stony Brook, NY, United States
| | - S Athar
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | - Z Yang
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | | | - D Samaras
- Department of Computer Science, Stony Brook University, Stony Brook, NY, United States
| | - J M Jarcho
- Department of Psychology, Temple University, Philadelphia, PA, United States.
| |
Collapse
|
5
|
Trujillo JP, Holler J. Interactionally Embedded Gestalt Principles of Multimodal Human Communication. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2023; 18:1136-1159. [PMID: 36634318 PMCID: PMC10475215 DOI: 10.1177/17456916221141422] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/14/2023]
Abstract
Natural human interaction requires us to produce and process many different signals, including speech, hand and head gestures, and facial expressions. These communicative signals, which occur in a variety of temporal relations with each other (e.g., parallel or temporally misaligned), must be rapidly processed as a coherent message by the receiver. In this contribution, we introduce the notion of interactionally embedded, affordance-driven gestalt perception as a framework that can explain how this rapid processing of multimodal signals is achieved as efficiently as it is. We discuss empirical evidence showing how basic principles of gestalt perception can explain some aspects of unimodal phenomena such as verbal language processing and visual scene perception but require additional features to explain multimodal human communication. We propose a framework in which high-level gestalt predictions are continuously updated by incoming sensory input, such as unfolding speech and visual signals. We outline the constituent processes that shape high-level gestalt perception and their role in perceiving relevance and prägnanz. Finally, we provide testable predictions that arise from this multimodal interactionally embedded gestalt-perception framework. This review and framework therefore provide a theoretically motivated account of how we may understand the highly complex, multimodal behaviors inherent in natural social interaction.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, the Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Nijmegen, the Netherlands
- Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| |
Collapse
|
6
|
Asalıoğlu EN, Göksun T. The role of hand gestures in emotion communication: Do type and size of gestures matter? PSYCHOLOGICAL RESEARCH 2023; 87:1880-1898. [PMID: 36436110 DOI: 10.1007/s00426-022-01774-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2022] [Accepted: 11/17/2022] [Indexed: 11/28/2022]
Abstract
We communicate emotions in a multimodal way, yet non-verbal emotion communication is a relatively understudied area of research. In three experiments, we investigated the role of gesture characteristics (e.g., type, size in space) on individuals' processing of emotional content. In Experiment 1, participants were asked to rate the emotional intensity of emotional narratives from the videoclips either with iconic or beat gestures. Participants in the iconic gesture condition rated the emotional intensity higher than participants in the beat gesture condition. In Experiment 2, the size of gestures and its interaction with gesture type were investigated in a within-subjects design. Participants again rated the emotional intensity of emotional narratives from the videoclips. Although individuals overall rated narrow gestures more emotionally intense than wider gestures, no effects of gesture type, or gesture size and type interaction were found. Experiment 3 was conducted to check whether findings of Experiment 2 were due to viewing gestures in all videoclips. We compared the gesture and no gesture (i.e., speech only) conditions and showed that there was not a difference between them on emotional ratings. However, we could not replicate the findings related to gesture size of Experiment 2. Overall, these findings indicate the importance of examining gesture's role in emotional contexts and that different gesture characteristics such as size of gestures can be considered in nonverbal communication.
Collapse
Affiliation(s)
- Esma Nur Asalıoğlu
- Department of Psychology, Koç University, Rumelifeneri Yolu, Sariyer, 34450, Istanbul, Turkey
| | - Tilbe Göksun
- Department of Psychology, Koç University, Rumelifeneri Yolu, Sariyer, 34450, Istanbul, Turkey.
| |
Collapse
|
7
|
Flaherty M, Sato A, Kirby S. Documenting a Reduction in Signing Space in Nicaraguan Sign Language Using Depth and Motion Capture. Cogn Sci 2023; 47:e13277. [PMID: 37096342 DOI: 10.1111/cogs.13277] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2020] [Revised: 01/27/2023] [Accepted: 03/06/2023] [Indexed: 04/26/2023]
Abstract
In this paper, we use motion tracking technology to document the birth of a brand new language: Nicaraguan Sign Language. Languages are dynamic entities that undergo change and growth through use, transmission, and learning, but the earliest stages of this process are generally difficult to observe as most languages have been used and passed down for many generations. Here, we observe a rare case of language emergence: the earliest stages of the new sign language in Nicaragua. By comparing the signing of the oldest and youngest signers of Nicaraguan Sign Language, we can track how the language itself is changing. Using motion tracking technology, we document a decrease in the size of articulatory space of Nicaraguan Sign Language signers over time. The reduction in articulatory space in Nicaraguan Sign appears to be the joint product of several decades of use and repeated transmission of this new language.
Collapse
Affiliation(s)
| | - Asha Sato
- Centre for Language Evolution, The University of Edinburgh
| | - Simon Kirby
- Centre for Language Evolution, The University of Edinburgh
| |
Collapse
|
8
|
Gestures and pauses to help thought: hands, voice, and silence in the tourist guide's speech. Cogn Process 2023; 24:25-41. [PMID: 36495353 DOI: 10.1007/s10339-022-01116-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2022] [Accepted: 11/23/2022] [Indexed: 12/14/2022]
Abstract
In the body of research on the relationship between gesture and speech, some models propose they form an integrated system while others attribute gestures a compensatory role in communication. This study addresses the gesture-speech relationship by taking disfluency phenomena as a case study. Since it is part of a project aimed at designing virtual agents to be employed in museums, an analysis was performed on the communicative behavior of tourist guides. Results reveal that gesturing is more frequent during speech than pauses. Moreover, when comparing the types of gestures and types of pauses they co-occur with, non-communicative gestures (idles and manipulators) turn out to be more frequent than communicatively-meaningful gestures, which instead more often co-occur with speech. We discuss these findings as relevant for a theoretical model viewing speech and gesture as an integrated system.
Collapse
|
9
|
Dockendorff M, Schmitz L, Vesper C, Knoblich G. Understanding others' distal goals from proximal communicative actions. PLoS One 2023; 18:e0280265. [PMID: 36662700 PMCID: PMC9858010 DOI: 10.1371/journal.pone.0280265] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Accepted: 12/23/2022] [Indexed: 01/21/2023] Open
Abstract
Many social interactions require individuals to coordinate their actions and to inform each other about their goals. Often these goals concern an immediate (i.e., proximal) action, as when people give each other a brief handshake, but they sometimes also refer to a future (i.e. distal) action, as when football players perform a passing sequence. The present study investigates whether observers can derive information about such distal goals by relying on kinematic modulations of an actor's instrumental actions. In Experiment 1 participants were presented with animations of a box being moved at different velocities towards an apparent endpoint. The distal goal, however, was for the object to be moved past this endpoint, to one of two occluded target locations. Participants then selected the location which they considered the likely distal goal of the action. As predicted, participants were able to detect differences in movement velocity and, based on these differences, systematically mapped the movements to the two distal goal locations. Adding a distal goal led to more variation in the way participants mapped the observed movements onto different target locations. The results of Experiments 2 and 3 indicated that this cannot be explained by difficulties in perceptual discrimination. Rather, the increased variability likely reflects differences in interpreting the underlying connection between proximal communicative actions and distal goals. The present findings extend previous research on sensorimotor communication by demonstrating that communicative action modulations are not restricted to predicting proximal goals but can also be used to infer more distal goals.
Collapse
Affiliation(s)
- Martin Dockendorff
- Department of Cognitive Science, Central European University, Vienna, Austria
| | - Laura Schmitz
- Department of Neurology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | - Cordula Vesper
- Department of Linguistics, Cognitive Science, and Semiotics, Aarhus University, Aarhus, Denmark
- Interacting Minds Centre, Aarhus University, Aarhus, Denmark
| | - Günther Knoblich
- Department of Cognitive Science, Central European University, Vienna, Austria
| |
Collapse
|
10
|
de’Sperati C, Dalmasso V, Moretti M, Høeg ER, Baud-Bovy G, Cozzi R, Ippolito J. Enhancing Visual Exploration through Augmented Gaze: High Acceptance of Immersive Virtual Biking by Oldest Olds. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2023; 20:1671. [PMID: 36767037 PMCID: PMC9914324 DOI: 10.3390/ijerph20031671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 01/12/2023] [Accepted: 01/13/2023] [Indexed: 06/18/2023]
Abstract
The diffusion of virtual reality applications dedicated to aging urges us to appraise its acceptance by target populations, especially the oldest olds. We investigated whether immersive virtual biking, and specifically a visuomotor manipulation aimed at improving visual exploration (augmented gaze), was well accepted by elders living in assisted residences. Twenty participants (mean age 89.8 years, five males) performed three 9 min virtual biking sessions pedalling on a cycle ergometer while wearing a Head-Mounted Display which immersed them inside a 360-degree pre-recorded biking video. In the second and third sessions, the relationship between horizontal head rotation and contingent visual shift was experimentally manipulated (augmented gaze), the visual shift being twice (gain = 2.0) or thrice (gain = 3.0) the amount of head rotation. User experience, motion sickness and visual exploration were measured. We found (i) very high user experience ratings, regardless of the gain; (ii) no effect of gain on motion sickness; and (iii) increased visual exploration (slope = +46%) and decreased head rotation (slope = -18%) with augmented gaze. The improvement in visual exploration capacity, coupled with the lack of intolerance signs, suggests that augmented gaze can be a valuable tool to improve the "visual usability" of certain virtual reality applications for elders, including the oldest olds.
Collapse
Affiliation(s)
- Claudio de’Sperati
- Laboratory of Action, Perception and Cognition, School of Psychology, Vita-Salute San Raffaele University, 20132 Milan, Italy
| | - Vittorio Dalmasso
- Laboratory of Action, Perception and Cognition, School of Psychology, Vita-Salute San Raffaele University, 20132 Milan, Italy
| | - Michela Moretti
- Laboratory of Action, Perception and Cognition, School of Psychology, Vita-Salute San Raffaele University, 20132 Milan, Italy
| | - Emil Rosenlund Høeg
- Multisensory Experience Laboratory, Department of Architecture, Design and Media Technology, Aalborg University, 2450 Copenhagen, Denmark
| | - Gabriel Baud-Bovy
- Laboratory of Action, Perception and Cognition, School of Psychology, Vita-Salute San Raffaele University, 20132 Milan, Italy
| | - Roberto Cozzi
- RSA San Giuseppe, Associazione Monte Tabor, 20132 Milan, Italy
| | - Jacopo Ippolito
- RSA San Giuseppe, Associazione Monte Tabor, 20132 Milan, Italy
| |
Collapse
|
11
|
Understanding conversational interaction in multiparty conversations: the EVA Corpus. LANG RESOUR EVAL 2022. [DOI: 10.1007/s10579-022-09627-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
AbstractThis paper focuses on gaining new knowledge through observation, qualitative analytics, and cross-modal fusion of rich multi-layered conversational features expressed during multiparty discourse. The outlined research stems from the theory that speech and co-speech gestures originate from the same representation; however, the representation is not solely limited to the speech production process. Thus, the nature of how information is conveyed by synchronously fusing speech and gestures must be investigated in detail. Therefore, this paper introduces an integrated annotation scheme and methodology which opens the opportunity to study verbal (i.e., speech) and non-verbal (i.e., visual cues with a communicative intent) components independently, however, still interconnected over a common timeline. To analyse this interaction between linguistic, paralinguistic, and non-verbal components in multiparty discourse and to help improve natural language generation in embodied conversational agents, a high-quality multimodal corpus, consisting of several annotation layers spanning syntax, POS, dialogue acts, discourse markers, sentiment, emotions, non-verbal behaviour, and gesture units was built and is represented in detail. It is the first of its kind for the Slovenian language. Moreover, detailed case studies show the tendency of metadiscourse to coincide with non-verbal behaviour of non-propositional origin. The case analysis further highlights how the newly created conversational model and the corresponding information-rich consistent corpus can be exploited to deepen the understanding of multiparty discourse.
Collapse
|
12
|
Eijk L, Rasenberg M, Arnese F, Blokpoel M, Dingemanse M, Doeller CF, Ernestus M, Holler J, Milivojevic B, Özyürek A, Pouw W, van Rooij I, Schriefers H, Toni I, Trujillo J, Bögels S. The CABB dataset: A multimodal corpus of communicative interactions for behavioural and neural analyses. Neuroimage 2022; 264:119734. [PMID: 36343884 DOI: 10.1016/j.neuroimage.2022.119734] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Revised: 10/07/2022] [Accepted: 11/03/2022] [Indexed: 11/06/2022] Open
Abstract
We present a dataset of behavioural and fMRI observations acquired in the context of humans involved in multimodal referential communication. The dataset contains audio/video and motion-tracking recordings of face-to-face, task-based communicative interactions in Dutch, as well as behavioural and neural correlates of participants' representations of dialogue referents. Seventy-one pairs of unacquainted participants performed two interleaved interactional tasks in which they described and located 16 novel geometrical objects (i.e., Fribbles) yielding spontaneous interactions of about one hour. We share high-quality video (from three cameras), audio (from head-mounted microphones), and motion-tracking (Kinect) data, as well as speech transcripts of the interactions. Before and after engaging in the face-to-face communicative interactions, participants' individual representations of the 16 Fribbles were estimated. Behaviourally, participants provided a written description (one to three words) for each Fribble and positioned them along 29 independent conceptual dimensions (e.g., rounded, human, audible). Neurally, fMRI signal evoked by each Fribble was measured during a one-back working-memory task. To enable functional hyperalignment across participants, the dataset also includes fMRI measurements obtained during visual presentation of eight animated movies (35 min total). We present analyses for the various types of data demonstrating their quality and consistency with earlier research. Besides high-resolution multimodal interactional data, this dataset includes different correlates of communicative referents, obtained before and after face-to-face dialogue, allowing for novel investigations into the relation between communicative behaviours and the representational space shared by communicators. This unique combination of data can be used for research in neuroscience, psychology, linguistics, and beyond.
Collapse
Affiliation(s)
- Lotte Eijk
- Centre for Language Studies, Radboud University, Nijmegen, the Netherlands
| | - Marlou Rasenberg
- Centre for Language Studies, Radboud University, Nijmegen, the Netherlands; Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Flavia Arnese
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - Mark Blokpoel
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - Mark Dingemanse
- Centre for Language Studies, Radboud University, Nijmegen, the Netherlands
| | - Christian F Doeller
- Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany; Kavli Institute for Systems Neuroscience, Centre for Neural Computation, The Egil and Pauline Braathen and Fred Kavli Centre for Cortical Microcircuits, Jebsen Centre for Alzheimer's Disease, Norwegian University of Science and Technology, Trondheim, Norway; Wilhelm Wundt Institute of Psychology, Leipzig University, Leipzig, Germany
| | - Mirjam Ernestus
- Centre for Language Studies, Radboud University, Nijmegen, the Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands; Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Branka Milivojevic
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - Asli Özyürek
- Centre for Language Studies, Radboud University, Nijmegen, the Netherlands; Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands; Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - Wim Pouw
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands; Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Iris van Rooij
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands; Department of Linguistics, Cognitive Science, and Semiotics, and the Interacting Minds Centre at Aarhus University, Denmark
| | - Herbert Schriefers
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - Ivan Toni
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands
| | - James Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands; Max Planck Institute for Psycholinguistics, Nijmegen, the Netherlands
| | - Sara Bögels
- Donders Institute for Brain, Cognition, and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, P.O.Box 9010, Nijmegen, Gelderland 6500, the Netherlands; Department of Cognition and Communication, Tilburg University, the Netherlands.
| |
Collapse
|
13
|
Močnik G, Kačič Z, Šafarič R, Mlakar I. Capturing Conversational Gestures for Embodied Conversational Agents Using an Optimized Kaneda-Lucas-Tomasi Tracker and Denavit-Hartenberg-Based Kinematic Model. SENSORS (BASEL, SWITZERLAND) 2022; 22:s22218318. [PMID: 36366016 PMCID: PMC9656321 DOI: 10.3390/s22218318] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/08/2022] [Revised: 10/17/2022] [Accepted: 10/27/2022] [Indexed: 05/27/2023]
Abstract
In order to recreate viable and human-like conversational responses, the artificial entity, i.e., an embodied conversational agent, must express correlated speech (verbal) and gestures (non-verbal) responses in spoken social interaction. Most of the existing frameworks focus on intent planning and behavior planning. The realization, however, is left to a limited set of static 3D representations of conversational expressions. In addition to functional and semantic synchrony between verbal and non-verbal signals, the final believability of the displayed expression is sculpted by the physical realization of non-verbal expressions. A major challenge of most conversational systems capable of reproducing gestures is the diversity in expressiveness. In this paper, we propose a method for capturing gestures automatically from videos and transforming them into 3D representations stored as part of the conversational agent's repository of motor skills. The main advantage of the proposed method is ensuring the naturalness of the embodied conversational agent's gestures, which results in a higher quality of human-computer interaction. The method is based on a Kanade-Lucas-Tomasi tracker, a Savitzky-Golay filter, a Denavit-Hartenberg-based kinematic model and the EVA framework. Furthermore, we designed an objective method based on cosine similarity instead of a subjective evaluation of synthesized movement. The proposed method resulted in a 96% similarity.
Collapse
|
14
|
Holler J. Visual bodily signals as core devices for coordinating minds in interaction. Philos Trans R Soc Lond B Biol Sci 2022; 377:20210094. [PMID: 35876208 PMCID: PMC9310176 DOI: 10.1098/rstb.2021.0094] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2021] [Accepted: 01/21/2022] [Indexed: 12/11/2022] Open
Abstract
The view put forward here is that visual bodily signals play a core role in human communication and the coordination of minds. Critically, this role goes far beyond referential and propositional meaning. The human communication system that we consider to be the explanandum in the evolution of language thus is not spoken language. It is, instead, a deeply multimodal, multilayered, multifunctional system that developed-and survived-owing to the extraordinary flexibility and adaptability that it endows us with. Beyond their undisputed iconic power, visual bodily signals (manual and head gestures, facial expressions, gaze, torso movements) fundamentally contribute to key pragmatic processes in modern human communication. This contribution becomes particularly evident with a focus that includes non-iconic manual signals, non-manual signals and signal combinations. Such a focus also needs to consider meaning encoded not just via iconic mappings, since kinematic modulations and interaction-bound meaning are additional properties equipping the body with striking pragmatic capacities. Some of these capacities, or its precursors, may have already been present in the last common ancestor we share with the great apes and may qualify as early versions of the components constituting the hypothesized interaction engine. This article is part of the theme issue 'Revisiting the human 'interaction engine': comparative approaches to social action coordination'.
Collapse
Affiliation(s)
- Judith Holler
- Max-Planck-Institut für Psycholinguistik, Nijmegen, The Netherlands
- Donders Centre for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| |
Collapse
|
15
|
Lozano-Goupil J, Raffard S, Capdevielle D, Aigoin E, Marin L. Gesture-speech synchrony in schizophrenia: A pilot study using a kinematic-acoustic analysis. Neuropsychologia 2022; 174:108347. [PMID: 35970254 DOI: 10.1016/j.neuropsychologia.2022.108347] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2022] [Revised: 07/30/2022] [Accepted: 08/08/2022] [Indexed: 10/15/2022]
Abstract
Severe impairment of social functioning is the core feature of schizophrenia that persists despite treatment, and contributes to chronic functional disability. Abnormal non-verbal behaviors have been reported during interpersonal interactions but the temporal coordination of co-speech gestures with language abilities have been poorly studied to date in this pathology. Using the dynamical systems framework, the goal of the current study was to investigate whether gestures and speech synchrony is impaired in schizophrenia, exploring a new approach to report communicational skill disorders. Performing the first continuous kinematic-acoustic analysis in individuals with schizophrenia, we examined gesture-speech synchrony in solo spontaneous speech and in sensorimotor synchronization task. The experimental group consisted of twenty-eight participants with a diagnosis of schizophrenia and the control group consisted of twenty-four healthy participants age-gender-education matched. The results showed that spontaneous gesture-speech synchrony was preserved while intentional finger tapping-speech synchrony was impaired. In sensorimotor synchronization task, the schizophrenia group displayed greater asynchronies between finger tapping and syllable uttering and lower stability of coordination patterns. These findings suggest a specific deficit in time delay of information circulation and processing, especially in explicit functions. Thus, investigating intrapersonal coordination in schizophrenia may constitute a promising window into brain/behavior dynamic relationship.
Collapse
Affiliation(s)
- Juliette Lozano-Goupil
- EuroMov Digital Health in Motion, Univ Montpellier, IMT Mines Ales, Montpellier, France.
| | - Stéphane Raffard
- Univ Paul Valéry Montpellier 3, EPSYLON EA, 4556, Montpellier, France; University Department of Adult Psychiatry, CHU Montpellier, Montpellier, France
| | | | - Emilie Aigoin
- EuroMov Digital Health in Motion, Univ Montpellier, IMT Mines Ales, Montpellier, France
| | - Ludovic Marin
- EuroMov Digital Health in Motion, Univ Montpellier, IMT Mines Ales, Montpellier, France
| |
Collapse
|
16
|
Trujillo JP, Özyürek A, Kan CC, Sheftel-Simanova I, Bekkering H. Differences in functional brain organization during gesture recognition between autistic and neurotypical individuals. Soc Cogn Affect Neurosci 2022; 17:1021-1034. [PMID: 35428885 PMCID: PMC9629468 DOI: 10.1093/scan/nsac026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2021] [Revised: 03/19/2022] [Accepted: 04/15/2022] [Indexed: 01/12/2023] Open
Abstract
Persons with and without autism process sensory information differently. Differences in sensory processing are directly relevant to social functioning and communicative abilities, which are known to be hampered in persons with autism. We collected functional magnetic resonance imaging data from 25 autistic individuals and 25 neurotypical individuals while they performed a silent gesture recognition task. We exploited brain network topology, a holistic quantification of how networks within the brain are organized to provide new insights into how visual communicative signals are processed in autistic and neurotypical individuals. Performing graph theoretical analysis, we calculated two network properties of the action observation network: 'local efficiency', as a measure of network segregation, and 'global efficiency', as a measure of network integration. We found that persons with autism and neurotypical persons differ in how the action observation network is organized. Persons with autism utilize a more clustered, local-processing-oriented network configuration (i.e. higher local efficiency) rather than the more integrative network organization seen in neurotypicals (i.e. higher global efficiency). These results shed new light on the complex interplay between social and sensory processing in autism.
Collapse
Affiliation(s)
- James P Trujillo
- Correspondence should be addressed to James P. Trujillo, Radboud University, Donders Centre for Cognition, Maria Montessori Building, Thomas van Aquinostraat 4, Nijmegen 6525 GD, The Netherlands. E-mail:
| | - Asli Özyürek
- Donders Institute for Brain, Cognition, and Behavior, Donders Centre for Cognition, Nijmegen, GD 6525, The Netherlands,Max Planck Institute for Psycholinguistics, Nijmegen, XD 6525, The Netherlands
| | - Cornelis C Kan
- Department of Psychiatry, Radboud University Medical Centre, Radboudumc, Nijmegen, GA 6525, The Netherlands
| | - Irina Sheftel-Simanova
- One Planet Research Centre, Radboud University Medical Centre, Radboudumc, Nijmegen, GA 6525, The Netherlands
| | - Harold Bekkering
- Donders Institute for Brain, Cognition, and Behavior, Donders Centre for Cognition, Nijmegen, GD 6525, The Netherlands
| |
Collapse
|
17
|
Trujillo JP, Levinson SC, Holler J. A multi-scale investigation of the human communication system's response to visual disruption. ROYAL SOCIETY OPEN SCIENCE 2022; 9:211489. [PMID: 35425638 PMCID: PMC9006025 DOI: 10.1098/rsos.211489] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/20/2021] [Accepted: 03/25/2022] [Indexed: 05/03/2023]
Abstract
In human communication, when the speech is disrupted, the visual channel (e.g. manual gestures) can compensate to ensure successful communication. Whether speech also compensates when the visual channel is disrupted is an open question, and one that significantly bears on the status of the gestural modality. We test whether gesture and speech are dynamically co-adapted to meet communicative needs. To this end, we parametrically reduce visibility during casual conversational interaction and measure the effects on speakers' communicative behaviour using motion tracking and manual annotation for kinematic and acoustic analyses. We found that visual signalling effort was flexibly adapted in response to a decrease in visual quality (especially motion energy, gesture rate, size, velocity and hold-time). Interestingly, speech was also affected: speech intensity increased in response to reduced visual quality (particularly in speech-gesture utterances, but independently of kinematics). Our findings highlight that multi-modal communicative behaviours are flexibly adapted at multiple scales of measurement and question the notion that gesture plays an inferior role to speech.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| | - Stephen C. Levinson
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD Nijmegen, The Netherlands
| |
Collapse
|
18
|
Zimmermann M, Lomoriello AS, Konvalinka I. Intra-individual behavioural and neural signatures of audience effects and interactions in a mirror-game paradigm. ROYAL SOCIETY OPEN SCIENCE 2022; 9:211352. [PMID: 35223056 PMCID: PMC8847899 DOI: 10.1098/rsos.211352] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/18/2021] [Accepted: 01/20/2022] [Indexed: 06/14/2023]
Abstract
We often perform actions while observed by others, yet the behavioural and neural signatures of audience effects remain understudied. Performing actions while being observed has been shown to result in more emphasized movements in musicians and dancers, as well as during communicative actions. Here, we investigate the behavioural and neural mechanisms of observed actions in relation to individual actions in isolation and interactive joint actions. Movement kinematics and EEG were recorded in 42 participants (21 pairs) during a mirror-game paradigm, while participants produced improvised movements alone, while observed by a partner, or by synchronizing movements with the partner. Participants produced largest movements when being observed, and observed actors and dyads in interaction produced slower and less variable movements in contrast with acting alone. On a neural level, we observed increased mu suppression during interaction, as well as to a lesser extent during observed actions, relative to individual actions. Moreover, we observed increased widespread functional brain connectivity during observed actions relative to both individual and interactive actions, suggesting increased intra-individual monitoring and action-perception integration as a result of audience effects. These results suggest that observed actors take observers into account in their action plans by increasing self-monitoring; on a behavioural level, observed actions are similar to emergent interactive actions, characterized by slower and more predictable movements.
Collapse
Affiliation(s)
- Marius Zimmermann
- Section for Cognitive Systems, DTU Compute, Kongens Lyngby, Denmark
- Institute of Psychology, University of Regensburg, Regensburg, Germany
| | | | - Ivana Konvalinka
- Section for Cognitive Systems, DTU Compute, Kongens Lyngby, Denmark
| |
Collapse
|
19
|
Fisher VJ. Embodied Songs: Insights Into the Nature of Cross-Modal Meaning-Making Within Sign Language Informed, Embodied Interpretations of Vocal Music. Front Psychol 2021; 12:624689. [PMID: 34744850 PMCID: PMC8569319 DOI: 10.3389/fpsyg.2021.624689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2020] [Accepted: 08/09/2021] [Indexed: 12/02/2022] Open
Abstract
Embodied song practices involve the transformation of songs from the acoustic modality into an embodied-visual form, to increase meaningful access for d/Deaf audiences. This goes beyond the translation of lyrics, by combining poetic sign language with other bodily movements to embody the para-linguistic expressive and musical features that enhance the message of a song. To date, the limited research into this phenomenon has focussed on linguistic features and interactions with rhythm. The relationship between bodily actions and music has not been probed beyond an assumed implication of conformance. However, as the primary objective is to communicate equivalent meanings, the ways that the acoustic and embodied-visual signals relate to each other should reveal something about underlying conceptual agreement. This paper draws together a range of pertinent theories from within a grounded cognition framework including semiotics, analogy mapping and cross-modal correspondences. These theories are applied to embodiment strategies used by prominent d/Deaf and hearing Dutch practitioners, to unpack the relationship between acoustic songs, their embodied representations, and their broader conceptual and affective meanings. This leads to the proposition that meaning primarily arises through shared patterns of internal relations across a range of amodal and cross-modal features with an emphasis on dynamic qualities. These analogous patterns can inform metaphorical interpretations and trigger shared emotional responses. This exploratory survey offers insights into the nature of cross-modal and embodied meaning-making, as a jumping-off point for further research.
Collapse
Affiliation(s)
- Vicky J. Fisher
- Multimodal Language and Cognition, Max Planck Institute for Pyscholinguistics and Centre for Language Studies, Radboud University, Nijmegen, Netherlands
| |
Collapse
|
20
|
Trujillo JP, Özyürek A, Kan CC, Sheftel-Simanova I, Bekkering H. Differences in the production and perception of communicative kinematics in autism. Autism Res 2021; 14:2640-2653. [PMID: 34536063 PMCID: PMC9292179 DOI: 10.1002/aur.2611] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 08/14/2021] [Accepted: 09/01/2021] [Indexed: 11/07/2022]
Abstract
In human communication, social intentions and meaning are often revealed in the way we move. In this study, we investigate the flexibility of human communication in terms of kinematic modulation in a clinical population, namely, autistic individuals. The aim of this study was twofold: to assess (a) whether communicatively relevant kinematic features of gestures differ between autistic and neurotypical individuals, and (b) if autistic individuals use communicative kinematic modulation to support gesture recognition. We tested autistic and neurotypical individuals on a silent gesture production task and a gesture comprehension task. We measured movement during the gesture production task using a Kinect motion tracking device in order to determine if autistic individuals differed from neurotypical individuals in their gesture kinematics. For the gesture comprehension task, we assessed whether autistic individuals used communicatively relevant kinematic cues to support recognition. This was done by using stick-light figures as stimuli and testing for a correlation between the kinematics of these videos and recognition performance. We found that (a) silent gestures produced by autistic and neurotypical individuals differ in communicatively relevant kinematic features, such as the number of meaningful holds between movements, and (b) while autistic individuals are overall unimpaired at recognizing gestures, they processed repetition and complexity, measured as the amount of submovements perceived, differently than neurotypicals do. These findings highlight how subtle aspects of neurotypical behavior can be experienced differently by autistic individuals. They further demonstrate the relationship between movement kinematics and social interaction in high-functioning autistic individuals. LAY SUMMARY: Hand gestures are an important part of how we communicate, and the way that we move when gesturing can influence how easy a gesture is to understand. We studied how autistic and typical individuals produce and recognize hand gestures, and how this relates to movement characteristics. We found that autistic individuals moved differently when gesturing compared to typical individuals. In addition, while autistic individuals were not worse at recognizing gestures, they differed from typical individuals in how they interpreted certain movement characteristics.
Collapse
Affiliation(s)
- James P Trujillo
- Donders Centre for Cognition, Donders Institute for Brain, Cognition, and Behavior, Nijmegen, The Netherlands.,Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Asli Özyürek
- Donders Centre for Cognition, Donders Institute for Brain, Cognition, and Behavior, Nijmegen, The Netherlands.,Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
| | - Cornelis C Kan
- Department of Psychiatry, Radboud University Medical Centre, Radboudumc, Nijmegen, The Netherlands
| | - Irina Sheftel-Simanova
- One Planet Research Centre, Radboud University Medical Centre, Radboudumc, Nijmegen, The Netherlands
| | - Harold Bekkering
- Donders Centre for Cognition, Donders Institute for Brain, Cognition, and Behavior, Nijmegen, The Netherlands
| |
Collapse
|
21
|
Trujillo J, Özyürek A, Holler J, Drijvers L. Speakers exhibit a multimodal Lombard effect in noise. Sci Rep 2021; 11:16721. [PMID: 34408178 PMCID: PMC8373897 DOI: 10.1038/s41598-021-95791-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Accepted: 07/29/2021] [Indexed: 12/03/2022] Open
Abstract
In everyday conversation, we are often challenged with communicating in non-ideal settings, such as in noise. Increased speech intensity and larger mouth movements are used to overcome noise in constrained settings (the Lombard effect). How we adapt to noise in face-to-face interaction, the natural environment of human language use, where manual gestures are ubiquitous, is currently unknown. We asked Dutch adults to wear headphones with varying levels of multi-talker babble while attempting to communicate action verbs to one another. Using quantitative motion capture and acoustic analyses, we found that (1) noise is associated with increased speech intensity and enhanced gesture kinematics and mouth movements, and (2) acoustic modulation only occurs when gestures are not present, while kinematic modulation occurs regardless of co-occurring speech. Thus, in face-to-face encounters the Lombard effect is not constrained to speech but is a multimodal phenomenon where the visual channel carries most of the communicative burden.
Collapse
Affiliation(s)
- James Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands.
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands.
| | - Asli Özyürek
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
- Centre for Language Studies, Radboud University Nijmegen, Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
| | - Linda Drijvers
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
| |
Collapse
|
22
|
Nota N, Trujillo JP, Holler J. Facial Signals and Social Actions in Multimodal Face-to-Face Interaction. Brain Sci 2021; 11:1017. [PMID: 34439636 PMCID: PMC8392358 DOI: 10.3390/brainsci11081017] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2021] [Revised: 07/07/2021] [Accepted: 07/26/2021] [Indexed: 01/30/2023] Open
Abstract
In a conversation, recognising the speaker's social action (e.g., a request) early may help the potential following speakers understand the intended message quickly, and plan a timely response. Human language is multimodal, and several studies have demonstrated the contribution of the body to communication. However, comparatively few studies have investigated (non-emotional) conversational facial signals and very little is known about how they contribute to the communication of social actions. Therefore, we investigated how facial signals map onto the expressions of two fundamental social actions in conversations: asking questions and providing responses. We studied the distribution and timing of 12 facial signals across 6778 questions and 4553 responses, annotated holistically in a corpus of 34 dyadic face-to-face Dutch conversations. Moreover, we analysed facial signal clustering to find out whether there are specific combinations of facial signals within questions or responses. Results showed a high proportion of facial signals, with a qualitatively different distribution in questions versus responses. Additionally, clusters of facial signals were identified. Most facial signals occurred early in the utterance, and had earlier onsets in questions. Thus, facial signals may critically contribute to the communication of social actions in conversation by providing social action-specific visual information.
Collapse
Affiliation(s)
- Naomi Nota
- Donders Institute for Brain, Cognition, and Behaviour, 6525 AJ Nijmegen, The Netherlands; (J.P.T.); (J.H.)
- Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, The Netherlands
| | - James P. Trujillo
- Donders Institute for Brain, Cognition, and Behaviour, 6525 AJ Nijmegen, The Netherlands; (J.P.T.); (J.H.)
- Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition, and Behaviour, 6525 AJ Nijmegen, The Netherlands; (J.P.T.); (J.H.)
- Max Planck Institute for Psycholinguistics, 6525 XD Nijmegen, The Netherlands
| |
Collapse
|
23
|
Trujillo JP, Holler J. The Kinematics of Social Action: Visual Signals Provide Cues for What Interlocutors Do in Conversation. Brain Sci 2021; 11:996. [PMID: 34439615 PMCID: PMC8393665 DOI: 10.3390/brainsci11080996] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2021] [Revised: 07/07/2021] [Accepted: 07/23/2021] [Indexed: 11/17/2022] Open
Abstract
During natural conversation, people must quickly understand the meaning of what the other speaker is saying. This concerns not just the semantic content of an utterance, but also the social action (i.e., what the utterance is doing-requesting information, offering, evaluating, checking mutual understanding, etc.) that the utterance is performing. The multimodal nature of human language raises the question of whether visual signals may contribute to the rapid processing of such social actions. However, while previous research has shown that how we move reveals the intentions underlying instrumental actions, we do not know whether the intentions underlying fine-grained social actions in conversation are also revealed in our bodily movements. Using a corpus of dyadic conversations combined with manual annotation and motion tracking, we analyzed the kinematics of the torso, head, and hands during the asking of questions. Manual annotation categorized these questions into six more fine-grained social action types (i.e., request for information, other-initiated repair, understanding check, stance or sentiment, self-directed, active participation). We demonstrate, for the first time, that the kinematics of the torso, head and hands differ between some of these different social action categories based on a 900 ms time window that captures movements starting slightly prior to or within 600 ms after utterance onset. These results provide novel insights into the extent to which our intentions shape the way that we move, and provide new avenues for understanding how this phenomenon may facilitate the fast communication of meaning in conversational interaction, social action, and conversation.
Collapse
Affiliation(s)
- James P. Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, 6525 GD Nijmegen, The Netherlands;
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525 XD Nijmegen, The Netherlands
| | - Judith Holler
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen, 6525 GD Nijmegen, The Netherlands;
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525 XD Nijmegen, The Netherlands
| |
Collapse
|
24
|
van der Meer HA, Sheftel-Simanova I, Kan CC, Trujillo JP. Translation, Cross-Cultural Adaptation, and Validation of a Dutch Version of the Actions and Feelings Questionnaire in Autistic and Neurotypical Adults. J Autism Dev Disord 2021; 52:1771-1777. [PMID: 34008098 PMCID: PMC8938389 DOI: 10.1007/s10803-021-05082-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/10/2021] [Indexed: 02/06/2023]
Abstract
The actions and feelings questionnaire (AFQ) provides a short, self-report measure of how well someone uses and understands visual communicative signals such as gestures. The objective of this study was to translate and cross-culturally adapt the AFQ into Dutch (AFQ-NL) and validate this new version in neurotypical and autistic populations. Translation and adaptation of the AFQ consisted of forward translation, synthesis, back translation, and expert review. In order to validate the AFQ-NL, we assessed convergent and divergent validity. We additionally assessed internal consistency using Cronbach’s alpha. Validation and reliability outcomes were all satisfactory. The AFQ-NL is a valid adaptation that can be used for both autistic and neurotypical populations in the Netherlands.
Collapse
Affiliation(s)
- Hedwig A van der Meer
- Academic Centre for Dentistry Amsterdam (ACTA), University of Amsterdam and Vrije Universiteit (VU) University Amsterdam, Department of Orofacial Pain and Dysfunction, Amsterdam, The Netherlands.,Center of Expertise Urban Vitality, Faculty of Health, Amsterdam University of Applied Sciences, Amsterdam, The Netherlands
| | - Irina Sheftel-Simanova
- One Planet Research Centre, Radboud University Medical Centre, Radboudumc, Nijmegen, The Netherlands
| | - Cornelis C Kan
- Department of Psychiatry, Radboud University Medical Centre, Radboudumc, Nijmegen, The Netherlands
| | - James P Trujillo
- Donders Institute for Brain, Cognition, and Behavior, Nijmegen, The Netherlands. .,Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands.
| |
Collapse
|
25
|
Cavallo A, Romeo L, Ansuini C, Battaglia F, Nobili L, Pontil M, Panzeri S, Becchio C. Identifying the signature of prospective motor control in children with autism. Sci Rep 2021; 11:3165. [PMID: 33542311 PMCID: PMC7862688 DOI: 10.1038/s41598-021-82374-2] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2020] [Accepted: 01/19/2021] [Indexed: 12/15/2022] Open
Abstract
Failure to develop prospective motor control has been proposed to be a core phenotypic marker of autism spectrum disorders (ASD). However, whether genuine differences in prospective motor control permit discriminating between ASD and non-ASD profiles over and above individual differences in motor output remains unclear. Here, we combined high precision measures of hand movement kinematics and rigorous machine learning analyses to determine the true power of prospective movement data to differentiate children with autism and typically developing children. Our results show that while movement is unique to each individual, variations in the kinematic patterning of sequential grasping movements genuinely differentiate children with autism from typically developing children. These findings provide quantitative evidence for a prospective motor control impairment in autism and indicate the potential to draw inferences about autism on the basis of movement kinematics.
Collapse
Affiliation(s)
- Andrea Cavallo
- Cognition, Motion and Neuroscience Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy.,Department of Psychology, Università degli Studi di Torino, Turin, Italy
| | - Luca Romeo
- Department of Information Engineering, Università Politecnica delle Marche, Ancona, Italy.,Computational Statistics and Machine Learning Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Caterina Ansuini
- Cognition, Motion and Neuroscience Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Francesca Battaglia
- Cognition, Motion and Neuroscience Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy.,Child Neuropsychiatric Unit, IRCCS Istituto G. Gaslini, Genoa, Italy
| | - Lino Nobili
- Child Neuropsychiatric Unit, IRCCS Istituto G. Gaslini, Genoa, Italy.,DINOGMI Department of Neurosciences, Rehabilitation, Ophthalmology, Genetics and Maternal and Children's Sciences, Università degli Studi di Genova, Genoa, Italy
| | - Massimiliano Pontil
- Computational Statistics and Machine Learning Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Stefano Panzeri
- Neural Computational Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Cristina Becchio
- Cognition, Motion and Neuroscience Laboratory, Center for Human Technologies, Istituto Italiano di Tecnologia, Genoa, Italy.
| |
Collapse
|
26
|
Özer D, Göksun T. Gesture Use and Processing: A Review on Individual Differences in Cognitive Resources. Front Psychol 2020; 11:573555. [PMID: 33250817 PMCID: PMC7674851 DOI: 10.3389/fpsyg.2020.573555] [Citation(s) in RCA: 25] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2020] [Accepted: 09/29/2020] [Indexed: 01/02/2023] Open
Abstract
Speakers use spontaneous hand gestures as they speak and think. These gestures serve many functions for speakers who produce them as well as for listeners who observe them. To date, studies in the gesture literature mostly focused on group-comparisons or the external sources of variation to examine when people use, process, and benefit from using and observing gestures. However, there are also internal sources of variation in gesture use and processing. People differ in how frequently they use gestures, how salient their gestures are, for what purposes they produce gestures, and how much they benefit from using and seeing gestures during comprehension and learning depending on their cognitive dispositions. This review addresses how individual differences in different cognitive skills relate to how people employ gestures in production and comprehension across different ages (from infancy through adulthood to healthy aging) from a functionalist perspective. We conclude that speakers and listeners can use gestures as a compensation tool during communication and thinking that interacts with individuals' cognitive dispositions.
Collapse
Affiliation(s)
- Demet Özer
- Department of Psychology, Koç University, Istanbul, Turkey
| | | |
Collapse
|
27
|
Trujillo JP, Simanova I, Bekkering H, Özyürek A. The communicative advantage: how kinematic signaling supports semantic comprehension. PSYCHOLOGICAL RESEARCH 2020; 84:1897-1911. [PMID: 31079227 PMCID: PMC7772160 DOI: 10.1007/s00426-019-01198-y] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2019] [Accepted: 05/02/2019] [Indexed: 11/04/2022]
Abstract
Humans are unique in their ability to communicate information through representational gestures which visually simulate an action (eg. moving hands as if opening a jar). Previous research indicates that the intention to communicate modulates the kinematics (e.g., velocity, size) of such gestures. If and how this modulation influences addressees' comprehension of gestures have not been investigated. Here we ask whether communicative kinematic modulation enhances semantic comprehension (i.e., identification) of gestures. We additionally investigate whether any comprehension advantage is due to enhanced early identification or late identification. Participants (n = 20) watched videos of representational gestures produced in a more- (n = 60) or less-communicative (n = 60) context and performed a forced-choice recognition task. We tested the isolated role of kinematics by removing visibility of actor's faces in Experiment I, and by reducing the stimuli to stick-light figures in Experiment II. Three video lengths were used to disentangle early identification from late identification. Accuracy and response time quantified main effects. Kinematic modulation was tested for correlations with task performance. We found higher gesture identification performance in more- compared to less-communicative gestures. However, early identification was only enhanced within a full visual context, while late identification occurred even when viewing isolated kinematics. Additionally, temporally segmented acts with more post-stroke holds were associated with higher accuracy. Our results demonstrate that communicative signaling, interacting with other visual cues, generally supports gesture identification, while kinematic modulation specifically enhances late identification in the absence of other cues. Results provide insights into mutual understanding processes as well as creating artificial communicative agents.
Collapse
Affiliation(s)
- James P Trujillo
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, B.01.25, 6525GR, Nijmegen, The Netherlands.
- Centre for Language Studies, Radboud University, Nijmegen, The Netherlands.
| | - Irina Simanova
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, B.01.25, 6525GR, Nijmegen, The Netherlands
| | - Harold Bekkering
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Montessorilaan 3, B.01.25, 6525GR, Nijmegen, The Netherlands
| | - Asli Özyürek
- Centre for Language Studies, Radboud University, Nijmegen, The Netherlands
- Max Planck Institute for Psycholinguistics, Wundtlaan 1, 6525XD, Nijmegen, The Netherlands
| |
Collapse
|
28
|
Trujillo JP, Simanova I, Özyürek A, Bekkering H. Seeing the Unexpected: How Brains Read Communicative Intent through Kinematics. Cereb Cortex 2020; 30:1056-1067. [PMID: 31504305 PMCID: PMC7132920 DOI: 10.1093/cercor/bhz148] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2019] [Revised: 06/06/2019] [Accepted: 06/12/2019] [Indexed: 11/12/2022] Open
Abstract
Social interaction requires us to recognize subtle cues in behavior, such as kinematic differences in actions and gestures produced with different social intentions. Neuroscientific studies indicate that the putative mirror neuron system (pMNS) in the premotor cortex and mentalizing system (MS) in the medial prefrontal cortex support inferences about contextually unusual actions. However, little is known regarding the brain dynamics of these systems when viewing communicatively exaggerated kinematics. In an event-related functional magnetic resonance imaging experiment, 28 participants viewed stick-light videos of pantomime gestures, recorded in a previous study, which contained varying degrees of communicative exaggeration. Participants made either social or nonsocial classifications of the videos. Using participant responses and pantomime kinematics, we modeled the probability of each video being classified as communicative. Interregion connectivity and activity were modulated by kinematic exaggeration, depending on the task. In the Social Task, communicativeness of the gesture increased activation of several pMNS and MS regions and modulated top-down coupling from the MS to the pMNS, but engagement of the pMNS and MS was not found in the nonsocial task. Our results suggest that expectation violations can be a key cue for inferring communicative intention, extending previous findings from wholly unexpected actions to more subtle social signaling.
Collapse
Affiliation(s)
- James P Trujillo
- Donders Institute for Brain, Cognition and Behaviour
- Centre for Language Studies, Radboud University Nijmegen, 6500HD Nijmegen, the Netherlands
| | | | - Asli Özyürek
- Centre for Language Studies, Radboud University Nijmegen, 6500HD Nijmegen, the Netherlands
- Max Planck Institute for Psycholinguistics, 6525XD Nijmegen, the Netherlands
| | | |
Collapse
|
29
|
Cañigueral R, Hamilton AFDC. The Role of Eye Gaze During Natural Social Interactions in Typical and Autistic People. Front Psychol 2019; 10:560. [PMID: 30930822 PMCID: PMC6428744 DOI: 10.3389/fpsyg.2019.00560] [Citation(s) in RCA: 47] [Impact Index Per Article: 9.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2018] [Accepted: 02/28/2019] [Indexed: 12/13/2022] Open
Abstract
Social interactions involve complex exchanges of a variety of social signals, such as gaze, facial expressions, speech and gestures. Focusing on the dual function of eye gaze, this review explores how the presence of an audience, communicative purpose and temporal dynamics of gaze allow interacting partners to achieve successful communication. First, we focus on how being watched modulates social cognition and behavior. We then show that the study of interpersonal gaze processing, particularly gaze temporal dynamics, can provide valuable understanding of social behavior in real interactions. We propose that the Interpersonal Gaze Processing model, which combines both sensing and signaling functions of eye gaze, provides a framework to make sense of gaze patterns in live interactions. Finally, we discuss how autistic individuals process the belief in being watched and interpersonal dynamics of gaze, and suggest that systematic manipulation of factors modulating gaze signaling can reveal which aspects of social eye gaze are challenging in autism.
Collapse
Affiliation(s)
- Roser Cañigueral
- Institute of Cognitive Neuroscience, Division of Psychology and Language Sciences, University College London, London, United Kingdom
| | | |
Collapse
|
30
|
Dockendorff M, Sebanz N, Knoblich G. Deviations from optimality should be an integral part of a working definition of SMC. Phys Life Rev 2019; 28:22-23. [DOI: 10.1016/j.plrev.2019.01.010] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2019] [Accepted: 01/24/2019] [Indexed: 10/27/2022]
|
31
|
Capozzi F, Latif N, Ristic J. It's not all in the face: reduced face visibility does not modulate social segmentation. VISUAL COGNITION 2019. [DOI: 10.1080/13506285.2019.1579774] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Affiliation(s)
| | - Nida Latif
- Department of Psychology, McGill University, Montreal, Canada
| | - Jelena Ristic
- Department of Psychology, McGill University, Montreal, Canada
| |
Collapse
|
32
|
Toward the markerless and automatic analysis of kinematic features: A toolkit for gesture and movement research. Behav Res Methods 2018; 51:769-777. [PMID: 30143970 PMCID: PMC6478643 DOI: 10.3758/s13428-018-1086-8] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
Abstract
Action, gesture, and sign represent unique aspects of human communication that use form and movement to convey meaning. Researchers typically use manual coding of video data to characterize naturalistic, meaningful movements at various levels of description, but the availability of markerless motion-tracking technology allows for quantification of the kinematic features of gestures or any meaningful human movement. We present a novel protocol for extracting a set of kinematic features from movements recorded with Microsoft Kinect. Our protocol captures spatial and temporal features, such as height, velocity, submovements/strokes, and holds. This approach is based on studies of communicative actions and gestures and attempts to capture features that are consistently implicated as important kinematic aspects of communication. We provide open-source code for the protocol, a description of how the features are calculated, a validation of these features as quantified by our protocol versus manual coders, and a discussion of how the protocol can be applied. The protocol effectively quantifies kinematic features that are important in the production (e.g., characterizing different contexts) as well as the comprehension (e.g., used by addressees to understand intent and semantics) of manual acts. The protocol can also be integrated with qualitative analysis, allowing fast and objective demarcation of movement units, providing accurate coding even of complex movements. This can be useful to clinicians, as well as to researchers studying multimodal communication or human–robot interactions. By making this protocol available, we hope to provide a tool that can be applied to understanding meaningful movement characteristics in human communication.
Collapse
|