1
|
Fajardo I, Gómez-Merino N, Ferrer A, Rodríguez-Ortiz IR. Hearing What You Can't See: Influence of Face Masks on Speech Perception and Eye Movement by Adults With Hearing Loss. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2024:1-21. [PMID: 39302873 DOI: 10.1044/2024_jslhr-22-00562] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/22/2024]
Abstract
PURPOSE The aim of the study was to analyze how face masks influence speech perception and time spent looking at the speaker's mouth and eyes by adults with and without hearing loss. METHOD Twenty participants with hearing loss and 20 without were asked to repeat Spanish words presented in various conditions, including different types of face masks (no mask, transparent window mask, and opaque mask FFP2) and presentation modes (audiovisual, video only, and audio only). Recognition accuracy and the percentage of time looking at the speaker's eyes and mouth (dwell time) were measured. RESULTS In the audiovisual condition, participants with hearing loss had significantly better word recognition scores when the speaker wore no mask compared to when they wore an opaque face mask. However, there were no differences between the transparent mask and no mask conditions. For those with typical hearing, the type of face mask did not affect speech recognition. Audiovisual presentation consistently improved speech recognition for participants with hearing loss across all face mask conditions, but for those with typical hearing, it only improved compared to video-only mode. These participants demonstrated a ceiling effect in audiovisual and audio-only modes. Regarding eye movement patterns, participants spent less time looking at the speaker's mouth and more time at the eyes when the speaker wore an opaque mask compared to no mask or a transparent mask. CONCLUSION The use of transparent face masks (ClearMask-type model) is recommended in contexts where face masks are still used (hospitals) to prevent the hindering effect of opaque masks (FFP2-type model) in speech perception among people with hearing loss, provided that any fogging of the window of the transparent mask is controlled by wiping it off as needed and the light is in front of the speaker to minimize shadows.
Collapse
Affiliation(s)
- Inmaculada Fajardo
- Departamento de Psicología Evolutiva y de la Educación and ERI-Lectura-Atypical Research Group, Universitat de València, Spain
- Red Lectin (Inclusive Reading Network: Network for Research and Innovation in Atypical Reading)
| | - Nadina Gómez-Merino
- Departamento de Psicología Evolutiva y de la Educación and ERI-Lectura-Atypical Research Group, Universitat de València, Spain
- Red Lectin (Inclusive Reading Network: Network for Research and Innovation in Atypical Reading)
| | - Antonio Ferrer
- Departamento de Psicología Evolutiva y de la Educación and ERI-Lectura-Atypical Research Group, Universitat de València, Spain
- Red Lectin (Inclusive Reading Network: Network for Research and Innovation in Atypical Reading)
| | - Isabel R Rodríguez-Ortiz
- Departamento de Psicología Evolutiva y de la Educación and Laboratorio de Diversidad, Cognición y Lenguaje, Universidad de Sevilla, Spain
- Red Lectin (Inclusive Reading Network: Network for Research and Innovation in Atypical Reading)
| |
Collapse
|
2
|
Deng X, McClay E, Jastrzebski E, Wang Y, Yeung HH. Visual scanning patterns of a talking face when evaluating phonetic information in a native and non-native language. PLoS One 2024; 19:e0304150. [PMID: 38805447 PMCID: PMC11132507 DOI: 10.1371/journal.pone.0304150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 05/07/2024] [Indexed: 05/30/2024] Open
Abstract
When comprehending speech, listeners can use information encoded in visual cues from a face to enhance auditory speech comprehension. For example, prior work has shown that the mouth movements reflect articulatory features of speech segments and durational information, while pitch and speech amplitude are primarily cued by eyebrow and head movements. Little is known about how the visual perception of segmental and prosodic speech information is influenced by linguistic experience. Using eye-tracking, we studied how perceivers' visual scanning of different regions on a talking face predicts accuracy in a task targeting both segmental versus prosodic information, and also asked how this was influenced by language familiarity. Twenty-four native English perceivers heard two audio sentences in either English or Mandarin (an unfamiliar, non-native language), which sometimes differed in segmental or prosodic information (or both). Perceivers then saw a silent video of a talking face, and judged whether that video matched either the first or second audio sentence (or whether both sentences were the same). First, increased looking to the mouth predicted correct responses only for non-native language trials. Second, the start of a successful search for speech information in the mouth area was significantly delayed in non-native versus native trials, but just when there were only prosodic differences in the auditory sentences, and not when there were segmental differences. Third, (in correct trials) the saccade amplitude in native language trials was significantly greater than in non-native trials, indicating more intensely focused fixations in the latter. Taken together, these results suggest that mouth-looking was generally more evident when processing a non-native versus native language in all analyses, but fascinatingly, when measuring perceivers' latency to fixate the mouth, this language effect was largest in trials where only prosodic information was useful for the task.
Collapse
Affiliation(s)
- Xizi Deng
- Department of Linguistics, Simon Fraser University, Burnaby BC, Canada
| | - Elise McClay
- Department of Linguistics, Simon Fraser University, Burnaby BC, Canada
| | - Erin Jastrzebski
- Department of Linguistics, Simon Fraser University, Burnaby BC, Canada
| | - Yue Wang
- Department of Linguistics, Simon Fraser University, Burnaby BC, Canada
| | - H. Henny Yeung
- Department of Linguistics, Simon Fraser University, Burnaby BC, Canada
| |
Collapse
|
3
|
Perez ND, Kleiman MJ, Barenholtz E. Visual fixations during processing of time-compressed audiovisual presentations. Atten Percept Psychophys 2024; 86:367-372. [PMID: 38175327 DOI: 10.3758/s13414-023-02838-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/19/2023] [Indexed: 01/05/2024]
Abstract
Time-compression is a technique that allows users to adjust the playback speed of audio recordings, but comprehension declines at higher speeds. Previous research has shown that under challenging auditory conditions people have a greater tendency to fixate regions closer to a speaker's mouth. In the current study, we investigated whether there is a similar tendency to fixate the mouth region for time-compressed stimuli. Participants were presented with a brief audiovisual lecture at different speeds, while eye fixations were recorded, and comprehension was tested. Results showed that the 50% compressed lecture group looked more at the nose compared to eye fixations for the normal lecture, and those in the 75% compressed group looked more towards the mouth. Greater compression decreased comprehension, but audiovisual information did not reduce this deficit. These results indicate that people seek out audiovisual information to overcome time-compression, demonstrating the flexibility of the multimodal attentional system.
Collapse
Affiliation(s)
- Nicole D Perez
- Division of Undergraduate Studies, Florida Atlantic University, 777 Glades Rd., Boca Raton, FL, 33433, USA.
| | - Michael J Kleiman
- Comprehensive Center for Brain Health, University of Miami Miller School of Medicine, Miami, FL, USA
| | - Elan Barenholtz
- Department of Psychology, Center for Complex Systems and Brain Sciences, Florida Atlantic University, Boca Raton, FL, USA
| |
Collapse
|
4
|
Tan SHJ, Kalashnikova M, Di Liberto GM, Crosse MJ, Burnham D. Seeing a Talking Face Matters: Gaze Behavior and the Auditory-Visual Speech Benefit in Adults' Cortical Tracking of Infant-directed Speech. J Cogn Neurosci 2023; 35:1741-1759. [PMID: 37677057 DOI: 10.1162/jocn_a_02044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/09/2023]
Abstract
In face-to-face conversations, listeners gather visual speech information from a speaker's talking face that enhances their perception of the incoming auditory speech signal. This auditory-visual (AV) speech benefit is evident even in quiet environments but is stronger in situations that require greater listening effort such as when the speech signal itself deviates from listeners' expectations. One example is infant-directed speech (IDS) presented to adults. IDS has exaggerated acoustic properties that are easily discriminable from adult-directed speech (ADS). Although IDS is a speech register that adults typically use with infants, no previous neurophysiological study has directly examined whether adult listeners process IDS differently from ADS. To address this, the current study simultaneously recorded EEG and eye-tracking data from adult participants as they were presented with auditory-only (AO), visual-only, and AV recordings of IDS and ADS. Eye-tracking data were recorded because looking behavior to the speaker's eyes and mouth modulates the extent of AV speech benefit experienced. Analyses of cortical tracking accuracy revealed that cortical tracking of the speech envelope was significant in AO and AV modalities for IDS and ADS. However, the AV speech benefit [i.e., AV > (A + V)] was only present for IDS trials. Gaze behavior analyses indicated differences in looking behavior during IDS and ADS trials. Surprisingly, looking behavior to the speaker's eyes and mouth was not correlated with cortical tracking accuracy. Additional exploratory analyses indicated that attention to the whole display was negatively correlated with cortical tracking accuracy of AO and visual-only trials in IDS. Our results underscore the nuances involved in the relationship between neurophysiological AV speech benefit and looking behavior.
Collapse
Affiliation(s)
- Sok Hui Jessica Tan
- The MARCS Institute of Brain, Behaviour and Development, Western Sydney University, Australia
- Science of Learning in Education Centre, Office of Education Research, National Institute of Education, Nanyang Technological University, Singapore
| | - Marina Kalashnikova
- The Basque Center on Cognition, Brain and Language
- IKERBASQUE, Basque Foundation for Science
| | - Giovanni M Di Liberto
- ADAPT Centre, School of Computer Science and Statistics, Trinity College Institute of Neuroscience, Trinity College, The University of Dublin, Ireland
| | - Michael J Crosse
- SEGOTIA, Galway, Ireland
- Trinity Center for Biomedical Engineering, Department of Mechanical, Manufacturing & Biomedical Engineering, Trinity College Dublin, Dublin, Ireland
| | - Denis Burnham
- The MARCS Institute of Brain, Behaviour and Development, Western Sydney University, Australia
| |
Collapse
|
5
|
Alviar C, Sahoo M, Edwards L, Jones W, Klin A, Lense M. Infant-directed song potentiates infants' selective attention to adults' mouths over the first year of life. Dev Sci 2023; 26:e13359. [PMID: 36527322 PMCID: PMC10276172 DOI: 10.1111/desc.13359] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Revised: 11/03/2022] [Accepted: 12/02/2022] [Indexed: 12/23/2022]
Abstract
The mechanisms by which infant-directed (ID) speech and song support language development in infancy are poorly understood, with most prior investigations focused on the auditory components of these signals. However, the visual components of ID communication are also of fundamental importance for language learning: over the first year of life, infants' visual attention to caregivers' faces during ID speech switches from a focus on the eyes to a focus on the mouth, which provides synchronous visual cues that support speech and language development. Caregivers' facial displays during ID song are highly effective for sustaining infants' attention. Here we investigate if ID song specifically enhances infants' attention to caregivers' mouths. 299 typically developing infants watched clips of female actors engaging them with ID song and speech longitudinally at six time points from 3 to 12 months of age while eye-tracking data was collected. Infants' mouth-looking significantly increased over the first year of life with a significantly greater increase during ID song versus speech. This difference was early-emerging (evident in the first 6 months of age) and sustained over the first year. Follow-up analyses indicated specific properties inherent to ID song (e.g., slower tempo, reduced rhythmic variability) in part contribute to infants' increased mouth-looking, with effects increasing with age. The exaggerated and expressive facial features that naturally accompany ID song may make it a particularly effective context for modulating infants' visual attention and supporting speech and language development in both typically developing infants and those with or at risk for communication challenges. A video abstract of this article can be viewed at https://youtu.be/SZ8xQW8h93A. RESEARCH HIGHLIGHTS: Infants' visual attention to adults' mouths during infant-directed speech has been found to support speech and language development. Infant-directed (ID) song promotes mouth-looking by infants to a greater extent than does ID speech across the first year of life. Features characteristic of ID song such as slower tempo, increased rhythmicity, increased audiovisual synchrony, and increased positive affect, all increase infants' attention to the mouth. The effects of song on infants' attention to the mouth are more prominent during the second half of the first year of life.
Collapse
Affiliation(s)
- Camila Alviar
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
| | - Manash Sahoo
- Marcus Autism Center, Children’s Healthcare of Atlanta, Atlanta, GA, USA
- Emory University School of Medicine, Atlanta, GA, USA
| | - Laura Edwards
- Marcus Autism Center, Children’s Healthcare of Atlanta, Atlanta, GA, USA
- Emory University School of Medicine, Atlanta, GA, USA
| | - Warren Jones
- Marcus Autism Center, Children’s Healthcare of Atlanta, Atlanta, GA, USA
- Emory University School of Medicine, Atlanta, GA, USA
| | - Ami Klin
- Marcus Autism Center, Children’s Healthcare of Atlanta, Atlanta, GA, USA
- Emory University School of Medicine, Atlanta, GA, USA
| | - Miriam Lense
- Department of Otolaryngology - Head & Neck Surgery, Vanderbilt University Medical Center, Nashville, TN, USA
- Vanderbilt Kennedy Center, Vanderbilt University Medical Center, Nashville, TN, USA
- The Curb Center for Art, Enterprise, and Public Policy, Vanderbilt University, Nashville, TN, USA
| |
Collapse
|
6
|
Viktorsson C, Valtakari NV, Falck-Ytter T, Hooge ITC, Rudling M, Hessels RS. Stable eye versus mouth preference in a live speech-processing task. Sci Rep 2023; 13:12878. [PMID: 37553414 PMCID: PMC10409748 DOI: 10.1038/s41598-023-40017-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2023] [Accepted: 08/03/2023] [Indexed: 08/10/2023] Open
Abstract
Looking at the mouth region is thought to be a useful strategy for speech-perception tasks. The tendency to look at the eyes versus the mouth of another person during speech processing has thus far mainly been studied using screen-based paradigms. In this study, we estimated the eye-mouth-index (EMI) of 38 adult participants in a live setting. Participants were seated across the table from an experimenter, who read sentences out loud for the participant to remember in both a familiar (English) and unfamiliar (Finnish) language. No statistically significant difference in the EMI between the familiar and the unfamiliar languages was observed. Total relative looking time at the mouth also did not predict the number of correctly identified sentences. Instead, we found that the EMI was higher during an instruction phase than during the speech-processing task. Moreover, we observed high intra-individual correlations in the EMI across the languages and different phases of the experiment. We conclude that there are stable individual differences in looking at the eyes versus the mouth of another person. Furthermore, this behavior appears to be flexible and dependent on the requirements of the situation (speech processing or not).
Collapse
Affiliation(s)
- Charlotte Viktorsson
- Development and Neurodiversity Lab, Department of Psychology, Uppsala University, Uppsala, Sweden.
| | - Niilo V Valtakari
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| | - Terje Falck-Ytter
- Development and Neurodiversity Lab, Department of Psychology, Uppsala University, Uppsala, Sweden
- Center of Neurodevelopmental Disorders (KIND), Division of Neuropsychiatry, Department of Women's and Children's Health, Karolinska Institutet, Stockholm, Sweden
| | - Ignace T C Hooge
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| | - Maja Rudling
- Development and Neurodiversity Lab, Department of Psychology, Uppsala University, Uppsala, Sweden
| | - Roy S Hessels
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| |
Collapse
|
7
|
Birulés J, Goupil L, Josse J, Fort M. The Role of Talking Faces in Infant Language Learning: Mind the Gap between Screen-Based Settings and Real-Life Communicative Interactions. Brain Sci 2023; 13:1167. [PMID: 37626523 PMCID: PMC10452843 DOI: 10.3390/brainsci13081167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Revised: 07/28/2023] [Accepted: 08/01/2023] [Indexed: 08/27/2023] Open
Abstract
Over the last few decades, developmental (psycho) linguists have demonstrated that perceiving talking faces audio-visually is important for early language acquisition. Using mostly well-controlled and screen-based laboratory approaches, this line of research has shown that paying attention to talking faces is likely to be one of the powerful strategies infants use to learn their native(s) language(s). In this review, we combine evidence from these screen-based studies with another line of research that has studied how infants learn novel words and deploy their visual attention during naturalistic play. In our view, this is an important step toward developing an integrated account of how infants effectively extract audiovisual information from talkers' faces during early language learning. We identify three factors that have been understudied so far, despite the fact that they are likely to have an important impact on how infants deploy their attention (or not) toward talking faces during social interactions: social contingency, speaker characteristics, and task- dependencies. Last, we propose ideas to address these issues in future research, with the aim of reducing the existing knowledge gap between current experimental studies and the many ways infants can and do effectively rely upon the audiovisual information extracted from talking faces in their real-life language environment.
Collapse
Affiliation(s)
- Joan Birulés
- Laboratoire de Psychologie et NeuroCognition, CNRS UMR 5105, Université Grenoble Alpes, 38058 Grenoble, France; (L.G.); (J.J.); (M.F.)
| | - Louise Goupil
- Laboratoire de Psychologie et NeuroCognition, CNRS UMR 5105, Université Grenoble Alpes, 38058 Grenoble, France; (L.G.); (J.J.); (M.F.)
| | - Jérémie Josse
- Laboratoire de Psychologie et NeuroCognition, CNRS UMR 5105, Université Grenoble Alpes, 38058 Grenoble, France; (L.G.); (J.J.); (M.F.)
| | - Mathilde Fort
- Laboratoire de Psychologie et NeuroCognition, CNRS UMR 5105, Université Grenoble Alpes, 38058 Grenoble, France; (L.G.); (J.J.); (M.F.)
- Centre de Recherche en Neurosciences de Lyon, INSERM U1028-CNRS UMR 5292, Université Lyon 1, 69500 Bron, France
| |
Collapse
|
8
|
Baron A, Harwood V, Kleinman D, Campanelli L, Molski J, Landi N, Irwin J. Where on the face do we look during phonemic restoration: An eye-tracking study. Front Psychol 2023; 14:1005186. [PMID: 37303890 PMCID: PMC10249372 DOI: 10.3389/fpsyg.2023.1005186] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2022] [Accepted: 04/28/2023] [Indexed: 06/13/2023] Open
Abstract
Face to face communication typically involves audio and visual components to the speech signal. To examine the effect of task demands on gaze patterns in response to a speaking face, adults participated in two eye-tracking experiments with an audiovisual (articulatory information from the mouth was visible) and a pixelated condition (articulatory information was not visible). Further, task demands were manipulated by having listeners respond in a passive (no response) or an active (button press response) context. The active experiment required participants to discriminate between speech stimuli and was designed to mimic environmental situations which require one to use visual information to disambiguate the speaker's message, simulating different listening conditions in real-world settings. Stimuli included a clear exemplar of the syllable /ba/ and a second exemplar in which the formant initial consonant was reduced creating an /a/-like consonant. Consistent with our hypothesis, results revealed that the greatest fixations to the mouth were present in the audiovisual active experiment and visual articulatory information led to a phonemic restoration effect for the /a/ speech token. In the pixelated condition, participants fixated on the eyes, and discrimination of the deviant token within the active experiment was significantly greater than the audiovisual condition. These results suggest that when required to disambiguate changes in speech, adults may look to the mouth for additional cues to support processing when it is available.
Collapse
Affiliation(s)
- Alisa Baron
- Department of Communicative Disorders, University of Rhode Island, Kingston, RI, United States
| | - Vanessa Harwood
- Department of Communicative Disorders, University of Rhode Island, Kingston, RI, United States
| | | | - Luca Campanelli
- Department of Communicative Disorders, The University of Alabama, Tuscaloosa, AL, United States
| | - Joseph Molski
- Department of Communicative Disorders, University of Rhode Island, Kingston, RI, United States
| | - Nicole Landi
- Haskins Laboratories, New Haven, CT, United States
- Department of Psychological Sciences, University of Connecticut, Storrs, CT, United States
| | - Julia Irwin
- Haskins Laboratories, New Haven, CT, United States
- Department of Psychology, Southern Connecticut State University, New Haven, CT, United States
| |
Collapse
|
9
|
Matthews N, Folivi F. Omit needless words: Sentence length perception. PLoS One 2023; 18:e0282146. [PMID: 36827285 PMCID: PMC9955962 DOI: 10.1371/journal.pone.0282146] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2022] [Accepted: 02/08/2023] [Indexed: 02/25/2023] Open
Abstract
Short sentences improve readability. Short sentences also promote social justice through accessibility and inclusiveness. Despite this, much remains unknown about sentence length perception-an important factor in producing readable writing. Accordingly, we conducted a psychophysical study using procedures from Signal Detection Theory to examine sentence length perception in naive adults. Participants viewed real-world full-page text samples and judged whether a bolded target sentence contained more or fewer than 17 words. The experiment yielded four findings. First, naïve adults perceived sentence length in real-world text samples quickly (median = 300-400 ms) and precisely (median = ~90% correct). Second, flipping real-world text samples upside-down generated no reaction-time cost and nearly no loss in the precision of sentence length perception. This differs from the large inversion effects that characterize other highly practiced, real-world perceptual tasks involving canonically oriented stimuli, most notably face perception and reading. Third, participants significantly underestimated the length of mirror-reversed sentences-but not upside-down, nor standard sentences. This finding parallels participants' familiarity with commonly occurring left-justified right-ragged text, and suggests a novel demonstration of left-lateralized anchoring in scene syntax. Fourth, error patterns demonstrated that participants achieved their high speed, high precision sentence-length judgments by heuristically counting text lines, not by explicitly counting words. This suggests practical advice for writing instructors to offer students. When copy editing, students can quickly and precisely identify their long sentences via a line-counting heuristic, e.g., "a 17-word sentence spans about 1.5 text lines". Students can subsequently improve a long sentence's readability and inclusiveness by omitting needless words.
Collapse
Affiliation(s)
- Nestor Matthews
- Department of Psychology, Denison University, Granville, OH, United States of America
| | - Folly Folivi
- Department of Psychology, Denison University, Granville, OH, United States of America
| |
Collapse
|
10
|
Looking at faces in the wild. Sci Rep 2023; 13:783. [PMID: 36646709 PMCID: PMC9842722 DOI: 10.1038/s41598-022-25268-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2022] [Accepted: 11/28/2022] [Indexed: 01/18/2023] Open
Abstract
Faces are key to everyday social interactions, but our understanding of social attention is based on experiments that present images of faces on computer screens. Advances in wearable eye-tracking devices now enable studies in unconstrained natural settings but this approach has been limited by manual coding of fixations. Here we introduce an automatic 'dynamic region of interest' approach that registers eye-fixations to bodies and faces seen while a participant moves through the environment. We show that just 14% of fixations are to faces of passersby, contrasting with prior screen-based studies that suggest faces automatically capture visual attention. We also demonstrate the potential for this new tool to help understand differences in individuals' social attention, and the content of their perceptual exposure to other people. Together, this can form the basis of a new paradigm for studying social attention 'in the wild' that opens new avenues for theoretical, applied and clinical research.
Collapse
|
11
|
Wever MC, van Houtum LAEM, Janssen LHC, Spruit IM, Tollenaar MS, aan het Rot M, Elzinga BM. Eyes on you: Ensuring empathic accuracy or signalling empathy? INTERNATIONAL JOURNAL OF PSYCHOLOGY 2022; 57:743-752. [PMID: 35698286 PMCID: PMC9796408 DOI: 10.1002/ijop.12862] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2022] [Accepted: 05/15/2022] [Indexed: 01/01/2023]
Abstract
The eye region is thought to play an important role in the ability to accurately infer others' feelings, or empathic accuracy (EA), which is an important skill for social interaction. However, most past studies used static pictures, including only visual information, and knowledge about the contribution of the eye region to EA when visual information is presented together with verbal content is lacking. We therefore examined whether eye gazing contributes to EA during videos of emotional autobiographical stories including both visual and verbal content. One hundred seven perceivers watched videos of targets talking about positive and negative life events and continuously rated the targets' feelings during the videos. Simultaneously, perceivers' eyes were tracked. After each video, perceivers reported on their feelings and the extent to which they empathized with and took the perspective of the targets. In contrast to studies using static pictures, we found that gazing to the eyes of targets during the videos did not significantly contribute to EA. At the same time, results on the association between the amount of gaze towards the eye region of targets and perceivers' state and trait empathy ratings suggest that eye gazing might signal empathy and social engagement to others.
Collapse
Affiliation(s)
- Mirjam C.M. Wever
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands,Leiden Institute for Brain and CognitionLeidenThe Netherlands
| | - Lisanne A. E. M. van Houtum
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands,Leiden Institute for Brain and CognitionLeidenThe Netherlands
| | - Loes H. C. Janssen
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands,Leiden Institute for Brain and CognitionLeidenThe Netherlands
| | - Iris M. Spruit
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands
| | - Marieke S. Tollenaar
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands,Leiden Institute for Brain and CognitionLeidenThe Netherlands
| | - Marije aan het Rot
- Department of PsychologyUniversity of GroningenGroningenThe Netherlands,School of Behavioural and Cognitive NeurosciencesUniversity of GroningenGroningenThe Netherlands
| | - Bernet M. Elzinga
- Institute of PsychologyLeiden UniversityLeidenThe Netherlands,Leiden Institute for Brain and CognitionLeidenThe Netherlands
| |
Collapse
|
12
|
The multisensory cocktail party problem in children: Synchrony-based segregation of multiple talking faces improves in early childhood. Cognition 2022; 228:105226. [PMID: 35882100 DOI: 10.1016/j.cognition.2022.105226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/13/2021] [Revised: 07/09/2022] [Accepted: 07/11/2022] [Indexed: 11/23/2022]
Abstract
Extraction of meaningful information from multiple talkers relies on perceptual segregation. The temporal synchrony statistics inherent in everyday audiovisual (AV) speech offer a powerful basis for perceptual segregation. We investigated the developmental emergence of synchrony-based perceptual segregation of multiple talkers in 3-7-year-old children. Children either saw four identical or four different faces articulating temporally jittered versions of the same utterance and heard the audible version of the same utterance either synchronized with one of the talkers or desynchronized with all of them. Eye tracking revealed that selective attention to the temporally synchronized talking face increased while attention to the desynchronized faces decreased with age and that attention to the talkers' mouth primarily drove responsiveness. These findings demonstrate that the temporal synchrony statistics inherent in fluent AV speech assume an increasingly greater role in perceptual segregation of the multisensory clutter created by multiple talking faces in early childhood.
Collapse
|
13
|
Linka M, Broda MD, Alsheimer T, de Haas B, Ramon M. Characteristic fixation biases in Super-Recognizers. J Vis 2022; 22:17. [PMID: 35900724 PMCID: PMC9344214 DOI: 10.1167/jov.22.8.17] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Accepted: 06/22/2022] [Indexed: 12/14/2022] Open
Abstract
Neurotypical observers show large and reliable individual differences in gaze behavior along several semantic object dimensions. Individual gaze behavior toward faces has been linked to face identity processing, including that of neurotypical observers. Here, we investigated potential gaze biases in Super-Recognizers (SRs), individuals with exceptional face identity processing skills. Ten SRs, identified with a novel conservative diagnostic framework, and 43 controls freely viewed 700 complex scenes depicting more than 5000 objects. First, we tested whether SRs and controls differ in fixation biases along four semantic dimensions: faces, text, objects being touched, and bodies. Second, we tested potential group differences in fixation biases toward eyes and mouths. Finally, we tested whether SRs fixate closer to the theoretical optimal fixation point for face identification. SRs showed a stronger gaze bias toward faces and away from text and touched objects, starting from the first fixation onward. Further, SRs spent a significantly smaller proportion of first fixations and dwell time toward faces on mouths but did not differ in dwell time or first fixations devoted to eyes. Face fixation of SRs also fell significantly closer to the theoretical optimal fixation point for identification, just below the eyes. Our findings suggest that reliable superiority for face identity processing is accompanied by early fixation biases toward faces and preferred saccadic landing positions close to the theoretical optimum for face identification. We discuss future directions to investigate the functional basis of individual fixation behavior and face identity processing ability.
Collapse
Affiliation(s)
- Marcel Linka
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| | | | - Tamara Alsheimer
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
- Applied Face Cognition Lab, University of Lausanne, Institute of Psychology, Lausanne, Switzerland
| | - Benjamin de Haas
- Experimental Psychology, Justus Liebig University Giessen, Giessen, Germany
| | - Meike Ramon
- Applied Face Cognition Lab, University of Lausanne, Institute of Psychology, Lausanne, Switzerland
| |
Collapse
|
14
|
Zhou HY, Yang HX, Wei Z, Wan GB, Lui SSY, Chan RCK. Audiovisual synchrony detection for fluent speech in early childhood: An eye-tracking study. Psych J 2022; 11:409-418. [PMID: 35350086 DOI: 10.1002/pchj.538] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Revised: 01/09/2022] [Accepted: 02/17/2022] [Indexed: 11/05/2022]
Abstract
During childhood, the ability to detect audiovisual synchrony gradually sharpens for simple stimuli such as flashbeeps and single syllables. However, little is known about how children perceive synchrony for natural and continuous speech. This study investigated young children's gaze patterns while they were watching movies of two identical speakers telling stories side by side. Only one speaker's lip movements matched the voices and the other one either led or lagged behind the soundtrack by 600 ms. Children aged 3-6 years (n = 94, 52.13% males) showed an overall preference for the synchronous speaker, with no age-related changes in synchrony-detection sensitivity as indicated by similar gaze patterns across ages. However, viewing time to the synchronous speech was significantly longer in the auditory-leading (AL) condition compared with that in the visual-leading (VL) condition, suggesting asymmetric sensitivities for AL versus VL asynchrony have already been established in early childhood. When further examining gaze patterns on dynamic faces, we found that more attention focused on the mouth region was an adaptive strategy to read visual speech signals and thus associated with increased viewing time of the synchronous videos. Attention to detail, one dimension of autistic traits featured by local processing, has been found to be correlated with worse performances in speech synchrony processing. These findings extended previous research by showing the development of speech synchrony perception in young children, and may have implications for clinical populations (e.g., autism) with impaired multisensory integration.
Collapse
Affiliation(s)
- Han-Yu Zhou
- Neuropsychology and Applied Cognitive Neuroscience Laboratory, CAS Key Laboratory of Mental Health, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of Chinese Academy of Sciences, Beijing, China
| | - Han-Xue Yang
- Neuropsychology and Applied Cognitive Neuroscience Laboratory, CAS Key Laboratory of Mental Health, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of Chinese Academy of Sciences, Beijing, China
| | - Zhen Wei
- Affiliated Shenzhen Maternity and Child Healthcare Hospital, Shenzhen, China
| | - Guo-Bin Wan
- Affiliated Shenzhen Maternity and Child Healthcare Hospital, Shenzhen, China
| | - Simon S Y Lui
- Department of Psychiatry, The University of Hong Kong, Hong Kong Special Administrative Region, China
| | - Raymond C K Chan
- Neuropsychology and Applied Cognitive Neuroscience Laboratory, CAS Key Laboratory of Mental Health, Institute of Psychology, Chinese Academy of Sciences, Beijing, China.,Department of Psychology, University of Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
15
|
Chengalur-Smith I, Nevo S, Fitzgerald B. Enhancing Hybrid OSS Development Through Agile Methods and High Media Synchronicity. DATA BASE FOR ADVANCES IN INFORMATION SYSTEMS 2021. [DOI: 10.1145/3508484.3508490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
Open source software (OSS) is increasingly being developed by hybrid teams that consist of a mix of company employees and volunteer developers. While hybrid OSS teams are becoming more prevalent, they also face unique challenges due to the involvement of different constituents. To address those challenges, this paper develops and validates a new organizing model. Specifically, the paper draws on media synchronicity theory (MST) to theorize that hybrid OSS teams would benefit from adopting an organizing model that involves practicing agile methods and using communication tools with multiple symbols sets and high transmission velocity. The paper also extends MST by conceptualizing the theory's key concept of communication convergence as consisting of two distinct dimensions: affective and cognitive convergence. Using primary survey data from hybrid OSS teams, the paper presents empirical evidence that such an organizing model can enhance those teams' affective convergence and cognitive convergence and, in turn, their development productivity and the quality of the software. In addition, the results show that affective convergence has a stronger impact on hybrid OSS teams' performance than cognitive convergence.
Collapse
|
16
|
Holleman GA, Hooge ITC, Huijding J, Deković M, Kemner C, Hessels RS. Gaze and speech behavior in parent–child interactions: The role of conflict and cooperation. CURRENT PSYCHOLOGY 2021. [DOI: 10.1007/s12144-021-02532-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
Abstract
AbstractA primary mode of human social behavior is face-to-face interaction. In this study, we investigated the characteristics of gaze and its relation to speech behavior during video-mediated face-to-face interactions between parents and their preadolescent children. 81 parent–child dyads engaged in conversations about cooperative and conflictive family topics. We used a dual-eye tracking setup that is capable of concurrently recording eye movements, frontal video, and audio from two conversational partners. Our results show that children spoke more in the cooperation-scenario whereas parents spoke more in the conflict-scenario. Parents gazed slightly more at the eyes of their children in the conflict-scenario compared to the cooperation-scenario. Both parents and children looked more at the other's mouth region while listening compared to while speaking. Results are discussed in terms of the role that parents and children take during cooperative and conflictive interactions and how gaze behavior may support and coordinate such interactions.
Collapse
|
17
|
Banks B, Gowen E, Munro KJ, Adank P. Eye Gaze and Perceptual Adaptation to Audiovisual Degraded Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:3432-3445. [PMID: 34463528 DOI: 10.1044/2021_jslhr-21-00106] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Purpose Visual cues from a speaker's face may benefit perceptual adaptation to degraded speech, but current evidence is limited. We aimed to replicate results from previous studies to establish the extent to which visual speech cues can lead to greater adaptation over time, extending existing results to a real-time adaptation paradigm (i.e., without a separate training period). A second aim was to investigate whether eye gaze patterns toward the speaker's mouth were related to better perception, hypothesizing that listeners who looked more at the speaker's mouth would show greater adaptation. Method A group of listeners (n = 30) was presented with 90 noise-vocoded sentences in audiovisual format, whereas a control group (n = 29) was presented with the audio signal only. Recognition accuracy was measured throughout and eye tracking was used to measure fixations toward the speaker's eyes and mouth in the audiovisual group. Results Previous studies were partially replicated: The audiovisual group had better recognition throughout and adapted slightly more rapidly, but both groups showed an equal amount of improvement overall. Longer fixations on the speaker's mouth in the audiovisual group were related to better overall accuracy. An exploratory analysis further demonstrated that the duration of fixations to the speaker's mouth decreased over time. Conclusions The results suggest that visual cues may not benefit adaptation to degraded speech as much as previously thought. Longer fixations on a speaker's mouth may play a role in successfully decoding visual speech cues; however, this will need to be confirmed in future research to fully understand how patterns of eye gaze are related to audiovisual speech recognition. All materials, data, and code are available at https://osf.io/2wqkf/.
Collapse
Affiliation(s)
- Briony Banks
- Division of Neuroscience and Experimental Psychology, Faculty of Biology, Medicine and Health, The University of Manchester, United Kingdom
| | - Emma Gowen
- Division of Neuroscience and Experimental Psychology, Faculty of Biology, Medicine and Health, The University of Manchester, United Kingdom
| | - Kevin J Munro
- Manchester Centre for Audiology and Deafness, Faculty of Biology, Medicine and Health, The University of Manchester, United Kingdom
- Manchester University NHS Foundation Trust, Manchester Academic Health Science Centre, United Kingdom
| | - Patti Adank
- Speech, Hearing and Phonetic Sciences, University College London, United Kingdom
| |
Collapse
|
18
|
Trotter AS, Banks B, Adank P. The Relevance of the Availability of Visual Speech Cues During Adaptation to Noise-Vocoded Speech. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2021; 64:2513-2528. [PMID: 34161748 DOI: 10.1044/2021_jslhr-20-00575] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Purpose This study first aimed to establish whether viewing specific parts of the speaker's face (eyes or mouth), compared to viewing the whole face, affected adaptation to distorted noise-vocoded sentences. Second, this study also aimed to replicate results on processing of distorted speech from lab-based experiments in an online setup. Method We monitored recognition accuracy online while participants were listening to noise-vocoded sentences. We first established if participants were able to perceive and adapt to audiovisual four-band noise-vocoded sentences when the entire moving face was visible (AV Full). Four further groups were then tested: a group in which participants viewed the moving lower part of the speaker's face (AV Mouth), a group in which participants only see the moving upper part of the face (AV Eyes), a group in which participants could not see the moving lower or upper face (AV Blocked), and a group in which participants saw an image of a still face (AV Still). Results Participants repeated around 40% of the key words correctly and adapted during the experiment, but only when the moving mouth was visible. In contrast, performance was at floor level, and no adaptation took place, in conditions when the moving mouth was occluded. Conclusions The results show the importance of being able to observe relevant visual speech information from the speaker's mouth region, but not the eyes/upper face region, when listening and adapting to distorted sentences online. Second, the results also demonstrated that it is feasible to run speech perception and adaptation studies online, but that not all findings reported for lab studies replicate. Supplemental Material https://doi.org/10.23641/asha.14810523.
Collapse
Affiliation(s)
- Antony S Trotter
- Speech, Hearing and Phonetic Sciences, University College London, United Kingdom
| | - Briony Banks
- Department of Psychology, Lancaster University, United Kingdom
| | - Patti Adank
- Speech, Hearing and Phonetic Sciences, University College London, United Kingdom
| |
Collapse
|
19
|
Frank K, Schuster L, Alfertshofer M, Baumbach SF, Herterich V, Giunta RE, Moellhoff N, Braig D, Ehrl D, Cotofana S. How Does Wearing a Facecover Influence the Eye Movement Pattern in Times of COVID-19? Aesthet Surg J 2021; 41:NP1118-NP1124. [PMID: 33693469 PMCID: PMC7989657 DOI: 10.1093/asj/sjab121] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022] Open
Abstract
Background Since the emergence of the COVID-19 pandemic facecovers have become a common sight. The effect of facecovers on the gaze when looking at faces has not been assessed yet. Objective The aim of the present study is to investigate a potential difference in eye movement pattern in observes which are exposed to images showing a face without and with facecover to identify if there is truly a change of gaze when identifying (masked) facial features. Materials and Methods The eye movement of a total of 64 study participants (28 males and 36 females) with a mean age of 31.84±9.0 years was analyzed in this cross-sectional observational study. Eye movement analysis was conducted based on positional changes of eye features within an x- and y- coordinate system while two images (face without/with facecover) were displayed for 8 seconds. Results The results of this study revealed that the sequence of focussing on facial regions was not altered when wearing a facecover and followed the sequence: perioral, nose, periorbital. Wearing a facecover significantly increased the time of focussing on the periorbital region and increased also the number of repeated eye fixations during the interval of visual stimulus presentation. No statistically significant differences were observed between male and female participants in their eye movement pattern across all investigated variables with p > 0.433. Conclusion Aesthetic practitioners could utilized the presented data and develop marketing and treatment strategies which majorly target the periorbital area understanding the altered eye movement pattern in times of COVID-19.
Collapse
Affiliation(s)
- Konstantin Frank
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Luca Schuster
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Michael Alfertshofer
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Sebastian Felix Baumbach
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Munich, Germany
| | - Viktoria Herterich
- Department of General, Trauma and Reconstructive Surgery, University Hospital, LMU Munich, Munich, Germany
| | - Riccardo E Giunta
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Nicholas Moellhoff
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - David Braig
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Denis Ehrl
- Department for Hand, Plastic and Aesthetic Surgery, Ludwig Maximilian University of Munich, Munich, Germany
| | - Sebastian Cotofana
- Department of Clinical Anatomy, Mayo Clinic College of Medicine and Science, Rochester, MN, USA
| |
Collapse
|
20
|
Lewkowicz DJ, Schmuckler M, Agrawal V. The multisensory cocktail party problem in adults: Perceptual segregation of talking faces on the basis of audiovisual temporal synchrony. Cognition 2021; 214:104743. [PMID: 33940250 DOI: 10.1016/j.cognition.2021.104743] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2020] [Revised: 04/16/2021] [Accepted: 04/21/2021] [Indexed: 10/21/2022]
Abstract
Social interactions often involve a cluttered multisensory scene consisting of multiple talking faces. We investigated whether audiovisual temporal synchrony can facilitate perceptual segregation of talking faces. Participants either saw four identical or four different talking faces producing temporally jittered versions of the same visible speech utterance and heard the audible version of the same speech utterance. The audible utterance was either synchronized with the visible utterance produced by one of the talking faces or not synchronized with any of them. Eye tracking indicated that participants exhibited a marked preference for the synchronized talking face, that they gazed more at the mouth than the eyes overall, that they gazed more at the eyes of an audiovisually synchronized than a desynchronized talking face, and that they gazed more at the mouth when all talking faces were audiovisually desynchronized. These findings demonstrate that audiovisual temporal synchrony plays a major role in perceptual segregation of multisensory clutter and that adults rely on differential scanning strategies of a talker's eyes and mouth to discover sources of multisensory coherence.
Collapse
Affiliation(s)
- David J Lewkowicz
- Haskins Laboratories, New Haven, CT, USA; Yale Child Study Center, New Haven, CT, USA.
| | - Mark Schmuckler
- Department of Psychology, University of Toronto at Scarborough, Toronto, Canada
| | | |
Collapse
|
21
|
Poncet F, Soussignan R, Jaffiol M, Gaudelus B, Leleu A, Demily C, Franck N, Baudouin JY. The spatial distribution of eye movements predicts the (false) recognition of emotional facial expressions. PLoS One 2021; 16:e0245777. [PMID: 33497409 PMCID: PMC7837501 DOI: 10.1371/journal.pone.0245777] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2020] [Accepted: 01/07/2021] [Indexed: 11/19/2022] Open
Abstract
Recognizing facial expressions of emotions is a fundamental ability for adaptation to the social environment. To date, it remains unclear whether the spatial distribution of eye movements predicts accurate recognition or, on the contrary, confusion in the recognition of facial emotions. In the present study, we asked participants to recognize facial emotions while monitoring their gaze behavior using eye-tracking technology. In Experiment 1a, 40 participants (20 women) performed a classic facial emotion recognition task with a 5-choice procedure (anger, disgust, fear, happiness, sadness). In Experiment 1b, a second group of 40 participants (20 women) was exposed to the same materials and procedure except that they were instructed to say whether (i.e., Yes/No response) the face expressed a specific emotion (e.g., anger), with the five emotion categories tested in distinct blocks. In Experiment 2, two groups of 32 participants performed the same task as in Experiment 1a while exposed to partial facial expressions composed of actions units (AUs) present or absent in some parts of the face (top, middle, or bottom). The coding of the AUs produced by the models showed complex facial configurations for most emotional expressions, with several AUs in common. Eye-tracking data indicated that relevant facial actions were actively gazed at by the decoders during both accurate recognition and errors. False recognition was mainly associated with the additional visual exploration of less relevant facial actions in regions containing ambiguous AUs or AUs relevant to other emotional expressions. Finally, the recognition of facial emotions from partial expressions showed that no single facial actions were necessary to effectively communicate an emotional state. In contrast, the recognition of facial emotions relied on the integration of a complex set of facial cues.
Collapse
Affiliation(s)
- Fanny Poncet
- Developmental Ethology and Cognitive Psychology Lab, Centre des Sciences du Goût et de l’Alimentation, AgroSup Dijon, CNRS, Inrae, Université Bourgogne Franche-Comté, Dijon, France
- * E-mail: (FP); (JYB)
| | - Robert Soussignan
- Developmental Ethology and Cognitive Psychology Lab, Centre des Sciences du Goût et de l’Alimentation, AgroSup Dijon, CNRS, Inrae, Université Bourgogne Franche-Comté, Dijon, France
| | - Margaux Jaffiol
- Developmental Ethology and Cognitive Psychology Lab, Centre des Sciences du Goût et de l’Alimentation, AgroSup Dijon, CNRS, Inrae, Université Bourgogne Franche-Comté, Dijon, France
| | - Baptiste Gaudelus
- Centre Ressource de Réhabilitation Psychosociale et de Remédiation Cognitive, Centre Hospitalier Le Vinatier & Université Lyon 1 (CNRS UMR 5229), Université de Lyon, Lyon, France
| | - Arnaud Leleu
- Developmental Ethology and Cognitive Psychology Lab, Centre des Sciences du Goût et de l’Alimentation, AgroSup Dijon, CNRS, Inrae, Université Bourgogne Franche-Comté, Dijon, France
| | - Caroline Demily
- Reference Center for Rare Diseases with Psychiatric Phenotype GénoPsy, Centre Hospitalier le Vinatier, Marc Jeannerod Institute (CNRS & Claude Bernard Lyon 1 University), Bron, France
| | - Nicolas Franck
- Centre Ressource de Réhabilitation Psychosociale et de Remédiation Cognitive, Centre Hospitalier Le Vinatier & Université Lyon 1 (CNRS UMR 5229), Université de Lyon, Lyon, France
| | - Jean-Yves Baudouin
- Laboratoire Développement, Individu, Processus, Handicap, Éducation (DIPHE), Département Psychologie du Développement, de l'Éducation et des Vulnérabilités (PsyDEV), Institut de Psychologie, Université de Lyon (Lumière Lyon 2), Lyon, France
- * E-mail: (FP); (JYB)
| |
Collapse
|
22
|
Vettori S, Van der Donck S, Nys J, Moors P, Van Wesemael T, Steyaert J, Rossion B, Dzhelyova M, Boets B. Combined frequency-tagging EEG and eye-tracking measures provide no support for the "excess mouth/diminished eye attention" hypothesis in autism. Mol Autism 2020; 11:94. [PMID: 33228763 PMCID: PMC7686749 DOI: 10.1186/s13229-020-00396-5] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2020] [Accepted: 11/02/2020] [Indexed: 12/20/2022] Open
Abstract
BACKGROUND Scanning faces is important for social interactions. Difficulty with the social use of eye contact constitutes one of the clinical symptoms of autism spectrum disorder (ASD). It has been suggested that individuals with ASD look less at the eyes and more at the mouth than typically developing (TD) individuals, possibly due to gaze aversion or gaze indifference. However, eye-tracking evidence for this hypothesis is mixed. While gaze patterns convey information about overt orienting processes, it is unclear how this is manifested at the neural level and how relative covert attention to the eyes and mouth of faces might be affected in ASD. METHODS We used frequency-tagging EEG in combination with eye tracking, while participants watched fast flickering faces for 1-min stimulation sequences. The upper and lower halves of the faces were presented at 6 Hz and 7.5 Hz or vice versa in different stimulation sequences, allowing to objectively disentangle the neural saliency of the eyes versus mouth region of a perceived face. We tested 21 boys with ASD (8-12 years old) and 21 TD control boys, matched for age and IQ. RESULTS Both groups looked longer at the eyes than the mouth, without any group difference in relative fixation duration to these features. TD boys looked significantly more to the nose, while the ASD boys looked more outside the face. EEG neural saliency data partly followed this pattern: neural responses to the upper or lower face half were not different between groups, but in the TD group, neural responses to the lower face halves were larger than responses to the upper part. Face exploration dynamics showed that TD individuals mostly maintained fixations within the same facial region, whereas individuals with ASD switched more often between the face parts. LIMITATIONS Replication in large and independent samples may be needed to validate exploratory results. CONCLUSIONS Combined eye-tracking and frequency-tagged neural responses show no support for the excess mouth/diminished eye gaze hypothesis in ASD. The more exploratory face scanning style observed in ASD might be related to their increased feature-based face processing style.
Collapse
Affiliation(s)
- Sofie Vettori
- Center for Developmental Psychiatry, Department of Neurosciences, University of Leuven (KU Leuven), Leuven, Belgium.
- Leuven Autism Research (LAuRes), University of Leuven (KU Leuven), Leuven, Belgium.
| | - Stephanie Van der Donck
- Center for Developmental Psychiatry, Department of Neurosciences, University of Leuven (KU Leuven), Leuven, Belgium
- Leuven Autism Research (LAuRes), University of Leuven (KU Leuven), Leuven, Belgium
| | - Jannes Nys
- Department of Physics and Astronomy, Ghent University, Ghent, Belgium
- IDLab - Department of Computer Science, University of Antwerp - IMEC, Antwerp, Belgium
| | - Pieter Moors
- Laboratory of Experimental Psychology, University of Leuven (KU Leuven), Leuven, Belgium
| | - Tim Van Wesemael
- Department of Electrical Engineering (ESAT), Stadius Center for Dynamical Systems, Signal Processing and Data Analytics, Leuven, Belgium
| | - Jean Steyaert
- Center for Developmental Psychiatry, Department of Neurosciences, University of Leuven (KU Leuven), Leuven, Belgium
- Leuven Autism Research (LAuRes), University of Leuven (KU Leuven), Leuven, Belgium
| | - Bruno Rossion
- Institute of Research in Psychological Science, Institute of Neuroscience, University of Louvain, Louvain-La-Neuve, Belgium
- CNRS, CRAN - UMR 7039, Université de Lorraine, 54000, Nancy, France
- CHRU-Nancy, Service de Neurologie, Université de Lorraine, 54000, Nancy, France
| | - Milena Dzhelyova
- Leuven Autism Research (LAuRes), University of Leuven (KU Leuven), Leuven, Belgium
- Institute of Research in Psychological Science, Institute of Neuroscience, University of Louvain, Louvain-La-Neuve, Belgium
| | - Bart Boets
- Center for Developmental Psychiatry, Department of Neurosciences, University of Leuven (KU Leuven), Leuven, Belgium
- Leuven Autism Research (LAuRes), University of Leuven (KU Leuven), Leuven, Belgium
| |
Collapse
|
23
|
Abstract
Gaze-where one looks, how long, and when-plays an essential part in human social behavior. While many aspects of social gaze have been reviewed, there is no comprehensive review or theoretical framework that describes how gaze to faces supports face-to-face interaction. In this review, I address the following questions: (1) When does gaze need to be allocated to a particular region of a face in order to provide the relevant information for successful interaction; (2) How do humans look at other people, and faces in particular, regardless of whether gaze needs to be directed at a particular region to acquire the relevant visual information; (3) How does gaze support the regulation of interaction? The work reviewed spans psychophysical research, observational research, and eye-tracking research in both lab-based and interactive contexts. Based on the literature overview, I sketch a framework for future research based on dynamic systems theory. The framework holds that gaze should be investigated in relation to sub-states of the interaction, encompassing sub-states of the interactors, the content of the interaction as well as the interactive context. The relevant sub-states for understanding gaze in interaction vary over different timescales from microgenesis to ontogenesis and phylogenesis. The framework has important implications for vision science, psychopathology, developmental science, and social robotics.
Collapse
Affiliation(s)
- Roy S Hessels
- Experimental Psychology, Helmholtz Institute, Utrecht University, Heidelberglaan 1, 3584CS, Utrecht, The Netherlands.
- Developmental Psychology, Heidelberglaan 1, 3584CS, Utrecht, The Netherlands.
| |
Collapse
|
24
|
Hessels RS, Benjamins JS, van Doorn AJ, Koenderink JJ, Holleman GA, Hooge ITC. Looking behavior and potential human interactions during locomotion. J Vis 2020; 20:5. [PMID: 33007079 PMCID: PMC7545070 DOI: 10.1167/jov.20.10.5] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
As humans move through parts of their environment, they meet others that may or may not try to interact with them. Where do people look when they meet others? We had participants wearing an eye tracker walk through a university building. On the way, they encountered nine “walkers.” Walkers were instructed to e.g. ignore the participant, greet him or her, or attempt to hand out a flyer. The participant's gaze was mostly directed to the currently relevant body parts of the walker. Thus, the participants gaze depended on the walker's action. Individual differences in participant's looking behavior were consistent across walkers. Participants who did not respond to the walker seemed to look less at that walker, although this difference was not statistically significant. We suggest that models of gaze allocation should take social motivation into account.
Collapse
Affiliation(s)
- Roy S Hessels
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands.,
| | - Jeroen S Benjamins
- Experimental Psychology, Helmholtz Institute, and Social, Health and Organizational Psychology, Utrecht University, Utrecht, the Netherlands.,
| | - Andrea J van Doorn
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands.,
| | - Jan J Koenderink
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands.,
| | - Gijs A Holleman
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands.,
| | - Ignace T C Hooge
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands.,
| |
Collapse
|
25
|
Lindblom J. A Radical Reassessment of the Body in Social Cognition. Front Psychol 2020; 11:987. [PMID: 32581915 PMCID: PMC7291370 DOI: 10.3389/fpsyg.2020.00987] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2019] [Accepted: 04/21/2020] [Indexed: 11/13/2022] Open
Abstract
The main issue addressed in this paper is to provide a reassessment of the role and relevance of the body in social cognition from a radical embodied cognitive science perspective. Initially, I provide a historical introduction of the traditional account of the body in cognitive science, which I here call the cognitivist view. I then present several lines of criticism raised against the cognitivist view advanced by more embodied, enacted and situated approaches in cognitive science, and related disciplines. Next, I analyze several approaches under the umbrella of embodied social cognition. My line of argument is that some of these approaches, although pointing toward the right direction of conceiving that the social mind is not merely contained inside the head, still fail to fully acknowledge the radically embodied social mind. I argue that the failure of these accounts of embodied social cognition could be associated with so-called 'simple embodiment.' The third part of this paper focuses on elaborating an alternative characterization of the radically embodied social mind that also tries to reduce the remaining problems with 'simple embodiment.' I draw upon two turns in radically embodied cognitive science, the enactive turn, and the intersubjective turn. On the one hand, there is the risk of focusing too much on the individual level in social cognition that may result in new kinds of methodological individualism that partly neglect the social dimension. On the other hand, socially distributed and socially extended approaches that pay more attention to the dynamics within social interaction may encounter the risk of ignoring the individual during social interaction dynamics and simultaneously not emphasizing the role of embodiment. The approach taken is to consider several ways of describing and incorporating the (individual) social mind at the social level that includes language. I outline some ideas and motivations for how to study and expand the field of radical embodied social cognition in the future, as well as pose the ubiquitous hazard of falling back into a cognitivism view in several ways.
Collapse
Affiliation(s)
- Jessica Lindblom
- Interaction Lab, School of Informatics, University of Skövde, Skövde, Sweden
| |
Collapse
|
26
|
Wegner-Clemens K, Rennig J, Beauchamp MS. A relationship between Autism-Spectrum Quotient and face viewing behavior in 98 participants. PLoS One 2020; 15:e0230866. [PMID: 32352984 PMCID: PMC7192493 DOI: 10.1371/journal.pone.0230866] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2019] [Accepted: 03/10/2020] [Indexed: 01/18/2023] Open
Abstract
Faces are one of the most important stimuli that we encounter, but humans vary dramatically in their behavior when viewing a face: some individuals preferentially fixate the eyes, others fixate the mouth, and still others show an intermediate pattern. The determinants of these large individual differences are unknown. However, individuals with Autism Spectrum Disorder (ASD) spend less time fixating the eyes of a viewed face than controls, suggesting the hypothesis that autistic traits in healthy adults might explain individual differences in face viewing behavior. Autistic traits were measured in 98 healthy adults recruited from an academic setting using the Autism-Spectrum Quotient, a validated 50-statement questionnaire. Fixations were measured using a video-based eye tracker while participants viewed two different types of audiovisual movies: short videos of talker speaking single syllables and longer videos of talkers speaking sentences in a social context. For both types of movies, there was a positive correlation between Autism-Spectrum Quotient score and percent of time fixating the lower half of the face that explained from 4% to 10% of the variance in individual face viewing behavior. This effect suggests that in healthy adults, autistic traits are one of many factors that contribute to individual differences in face viewing behavior.
Collapse
Affiliation(s)
- Kira Wegner-Clemens
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, Texas, United States of America
| | - Johannes Rennig
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, Texas, United States of America
| | - Michael S. Beauchamp
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, Texas, United States of America
| |
Collapse
|
27
|
Holleman GA, Hessels RS, Kemner C, Hooge ITC. Implying social interaction and its influence on gaze behavior to the eyes. PLoS One 2020; 15:e0229203. [PMID: 32092089 PMCID: PMC7039466 DOI: 10.1371/journal.pone.0229203] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2019] [Accepted: 01/31/2020] [Indexed: 11/18/2022] Open
Abstract
Researchers have increasingly focused on how the potential for social interaction modulates basic processes of visual attention and gaze behavior. In this study, we investigated why people may experience social interaction and what factors contributed to their subjective experience. We furthermore investigated whether implying social interaction modulated gaze behavior to people’s faces, specifically the eyes. To imply the potential for interaction, participants received either one of two instructions: 1) they would be presented with a person via a ‘live’ video-feed, or 2) they would be presented with a pre-recorded video clip of a person. Prior to the presentation, a confederate walked into a separate room to suggest to participants that (s)he was being positioned behind a webcam. In fact, all participants were presented with a pre-recorded clip. During the presentation, we measured participants’ gaze behavior with an eye tracker, and after the presentation, participants were asked whether they believed that the confederate was ‘live’ or not, and, why they thought so. Participants varied greatly in their judgements about whether the confederate was ‘live’ or not. Analyses of gaze behavior revealed that a large subset of participants who received the live-instruction gazed less at the eyes of confederates compared with participants who received the pre-recorded-instruction. However, for both the live-instruction group and the pre-recorded instruction group, another subset of participants gazed predominantly at the eyes. The current findings may contribute to the development of experimental designs aimed to capture the interactive aspects of social cognition and visual attention.
Collapse
Affiliation(s)
- Gijs A. Holleman
- Experimental psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands
- Developmental psychology, Utrecht University, Utrecht, the Netherlands
- * E-mail:
| | - Roy S. Hessels
- Experimental psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands
- Developmental psychology, Utrecht University, Utrecht, the Netherlands
| | - Chantal Kemner
- Experimental psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands
- Developmental psychology, Utrecht University, Utrecht, the Netherlands
- Brain Center, University Medical Center Utrecht, Utrecht, the Netherlands
| | - Ignace T. C. Hooge
- Experimental psychology, Helmholtz Institute, Utrecht University, Utrecht, the Netherlands
| |
Collapse
|
28
|
Abstract
Previous research has shown that gaze behavior of a speaker's face during speech encoding is influenced by an array of factors relating to the quality of the speech signal and the encoding task. In these studies, participants were aware they were viewing pre-recorded stimuli of a speaker that is not representative of natural social interactions in which an interlocutor can observe one's gaze direction, potentially affecting fixation behavior due to communicative and social considerations. To assess the potential role of these factors during speech encoding, we compared fixation behavior during a speech-encoding task under two conditions: in the "real-time" condition, we used deception to convince participants that they were interacting with a live person who was able to see and hear them through online remote video communication. In the "pre-recorded" condition, participants were correctly informed they were watching a previously recorded video. We found that participants fixated the interlocutor's face significantly less in the real-time condition than the pre-recorded condition. When participants did look at the face, they fixated the mouth at a higher proportion of the time in the pre-recorded condition versus the real-time condition. These findings suggest that people engage in avoidance of potentially useful speech-directed fixations when they believe their fixations are being observed and demonstrate that social factors play a significant role in fixation behavior during speech encoding.
Collapse
|
29
|
Correia-Caeiro C, Guo K, Mills DS. Perception of dynamic facial expressions of emotion between dogs and humans. Anim Cogn 2020; 23:465-476. [PMID: 32052285 PMCID: PMC7181561 DOI: 10.1007/s10071-020-01348-5] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2019] [Revised: 01/07/2020] [Accepted: 01/14/2020] [Indexed: 11/29/2022]
Abstract
Facial expressions are a core component of the emotional response of social mammals. In contrast to Darwin's original proposition, expressive facial cues of emotion appear to have evolved to be species-specific. Faces trigger an automatic perceptual process, and so, inter-specific emotion perception is potentially a challenge; since observers should not try to “read” heterospecific facial expressions in the same way that they do conspecific ones. Using dynamic spontaneous facial expression stimuli, we report the first inter-species eye-tracking study on fully unrestrained participants and without pre-experiment training to maintain attention to stimuli, to compare how two different species living in the same ecological niche, humans and dogs, perceive each other’s facial expressions of emotion. Humans and dogs showed different gaze distributions when viewing the same facial expressions of either humans or dogs. Humans modulated their gaze depending on the area of interest (AOI) being examined, emotion, and species observed, but dogs modulated their gaze depending on AOI only. We also analysed if the gaze distribution was random across AOIs in both species: in humans, eye movements were not correlated with the diagnostic facial movements occurring in the emotional expression, and in dogs, there was only a partial relationship. This suggests that the scanning of facial expressions is a relatively automatic process. Thus, to read other species’ facial emotions successfully, individuals must overcome these automatic perceptual processes and employ learning strategies to appreciate the inter-species emotional repertoire.
Collapse
Affiliation(s)
- Catia Correia-Caeiro
- School of Psychology, University of Lincoln, Lincoln, UK. .,School of Life Sciences, University of Lincoln, Lincoln, UK.
| | - Kun Guo
- School of Psychology, University of Lincoln, Lincoln, UK
| | - Daniel S Mills
- School of Life Sciences, University of Lincoln, Lincoln, UK
| |
Collapse
|
30
|
Wegner-Clemens K, Rennig J, Magnotti JF, Beauchamp MS. Using principal component analysis to characterize eye movement fixation patterns during face viewing. J Vis 2019; 19:2. [PMID: 31689715 PMCID: PMC6833982 DOI: 10.1167/19.13.2] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2019] [Accepted: 08/23/2019] [Indexed: 01/22/2023] Open
Abstract
Human faces contain dozens of visual features, but viewers preferentially fixate just two of them: the eyes and the mouth. Face-viewing behavior is usually studied by manually drawing regions of interest (ROIs) on the eyes, mouth, and other facial features. ROI analyses are problematic as they require arbitrary experimenter decisions about the location and number of ROIs, and they discard data because all fixations within each ROI are treated identically and fixations outside of any ROI are ignored. We introduce a data-driven method that uses principal component analysis (PCA) to characterize human face-viewing behavior. All fixations are entered into a PCA, and the resulting eigenimages provide a quantitative measure of variability in face-viewing behavior. In fixation data from 41 participants viewing four face exemplars under three stimulus and task conditions, the first principal component (PC1) separated the eye and mouth regions of the face. PC1 scores varied widely across participants, revealing large individual differences in preference for eye or mouth fixation, and PC1 scores varied by condition, revealing the importance of behavioral task in determining fixation location. Linear mixed effects modeling of the PC1 scores demonstrated that task condition accounted for 41% of the variance, individual differences accounted for 28% of the variance, and stimulus exemplar for less than 1% of the variance. Fixation eigenimages provide a useful tool for investigating the relative importance of the different factors that drive human face-viewing behavior.
Collapse
Affiliation(s)
- Kira Wegner-Clemens
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, TX
| | - Johannes Rennig
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, TX
| | - John F Magnotti
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, TX
| | - Michael S Beauchamp
- Department of Neurosurgery and Core for Advanced MRI, Baylor College of Medicine, Houston, TX
| |
Collapse
|
31
|
Drijvers L, Vaitonytė J, Özyürek A. Degree of Language Experience Modulates Visual Attention to Visible Speech and Iconic Gestures During Clear and Degraded Speech Comprehension. Cogn Sci 2019; 43:e12789. [PMID: 31621126 PMCID: PMC6790953 DOI: 10.1111/cogs.12789] [Citation(s) in RCA: 21] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2018] [Revised: 07/12/2019] [Accepted: 08/19/2019] [Indexed: 11/27/2022]
Abstract
Visual information conveyed by iconic hand gestures and visible speech can enhance speech comprehension under adverse listening conditions for both native and non-native listeners. However, how a listener allocates visual attention to these articulators during speech comprehension is unknown. We used eye-tracking to investigate whether and how native and highly proficient non-native listeners of Dutch allocated overt eye gaze to visible speech and gestures during clear and degraded speech comprehension. Participants watched video clips of an actress uttering a clear or degraded (6-band noise-vocoded) action verb while performing a gesture or not, and were asked to indicate the word they heard in a cued-recall task. Gestural enhancement was the largest (i.e., a relative reduction in reaction time cost) when speech was degraded for all listeners, but it was stronger for native listeners. Both native and non-native listeners mostly gazed at the face during comprehension, but non-native listeners gazed more often at gestures than native listeners. However, only native but not non-native listeners' gaze allocation to gestures predicted gestural benefit during degraded speech comprehension. We conclude that non-native listeners might gaze at gesture more as it might be more challenging for non-native listeners to resolve the degraded auditory cues and couple those cues to phonological information that is conveyed by visible speech. This diminished phonological knowledge might hinder the use of semantic information that is conveyed by gestures for non-native compared to native listeners. Our results demonstrate that the degree of language experience impacts overt visual attention to visual articulators, resulting in different visual benefits for native versus non-native listeners.
Collapse
Affiliation(s)
- Linda Drijvers
- Donders Institute for Brain, Cognition, and BehaviourRadboud University
| | - Julija Vaitonytė
- Department of Cognitive and Artificial Intelligence (School of Humanities and Digital Sciences)Tilburg University
| | - Asli Özyürek
- Donders Institute for Brain, Cognition, and BehaviourRadboud University
- Centre for Language StudiesRadboud University
- Max Planck Institute for Psycholinguistics
| |
Collapse
|
32
|
Plouffe-Demers MP, Fiset D, Saumure C, Duncan J, Blais C. Strategy Shift Toward Lower Spatial Frequencies in the Recognition of Dynamic Facial Expressions of Basic Emotions: When It Moves It Is Different. Front Psychol 2019; 10:1563. [PMID: 31379648 PMCID: PMC6650765 DOI: 10.3389/fpsyg.2019.01563] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2018] [Accepted: 06/20/2019] [Indexed: 11/15/2022] Open
Abstract
Facial expressions of emotion play a key role in social interactions. While in everyday life, their dynamic and transient nature calls for a fast processing of the visual information they contain, a majority of studies investigating the visual processes underlying their recognition have focused on their static display. The present study aimed to gain a better understanding of these processes while using more ecological dynamic facial expressions. In two experiments, we directly compared the spatial frequency (SF) tuning during the recognition of static and dynamic facial expressions. Experiment 1 revealed a shift toward lower SFs for dynamic expressions in comparison to static ones. Experiment 2 was designed to verify if changes in SF tuning curves were specific to the presence of emotional information in motion by comparing the SF tuning profiles for static, dynamic, and shuffled dynamic expressions. Results showed a similar shift toward lower SFs for shuffled expressions, suggesting that the difference found between dynamic and static expressions might not be linked to informative motion per se but to the presence of motion regardless its nature.
Collapse
Affiliation(s)
- Marie-Pier Plouffe-Demers
- Département de Psychologie, Universtité du Québec en Outaouais, Gatineau, QC, Canada
- Département de Psychologie, Université du Québec à Montréal, Montreal, QC, Canada
| | - Daniel Fiset
- Département de Psychologie, Universtité du Québec en Outaouais, Gatineau, QC, Canada
| | - Camille Saumure
- Département de Psychologie, Universtité du Québec en Outaouais, Gatineau, QC, Canada
| | - Justin Duncan
- Département de Psychologie, Universtité du Québec en Outaouais, Gatineau, QC, Canada
- Département de Psychologie, Université du Québec à Montréal, Montreal, QC, Canada
| | - Caroline Blais
- Département de Psychologie, Universtité du Québec en Outaouais, Gatineau, QC, Canada
| |
Collapse
|
33
|
Psychobiological Responses Reveal Audiovisual Noise Differentially Challenges Speech Recognition. Ear Hear 2019; 41:268-277. [PMID: 31283529 DOI: 10.1097/aud.0000000000000755] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES In noisy environments, listeners benefit from both hearing and seeing a talker, demonstrating audiovisual (AV) cues enhance speech-in-noise (SIN) recognition. Here, we examined the relative contribution of auditory and visual cues to SIN perception and the strategies used by listeners to decipher speech in noise interference(s). DESIGN Normal-hearing listeners (n = 22) performed an open-set speech recognition task while viewing audiovisual TIMIT sentences presented under different combinations of signal degradation including visual (AVn), audio (AnV), or multimodal (AnVn) noise. Acoustic and visual noises were matched in physical signal-to-noise ratio. Eyetracking monitored participants' gaze to different parts of a talker's face during SIN perception. RESULTS As expected, behavioral performance for clean sentence recognition was better for A-only and AV compared to V-only speech. Similarly, with noise in the auditory channel (AnV and AnVn speech), performance was aided by the addition of visual cues of the talker regardless of whether the visual channel contained noise, confirming a multimodal benefit to SIN recognition. The addition of visual noise (AVn) obscuring the talker's face had little effect on speech recognition by itself. Listeners' eye gaze fixations were biased toward the eyes (decreased at the mouth) whenever the auditory channel was compromised. Fixating on the eyes was negatively associated with SIN recognition performance. Eye gazes on the mouth versus eyes of the face also depended on the gender of the talker. CONCLUSIONS Collectively, results suggest listeners (1) depend heavily on the auditory over visual channel when seeing and hearing speech and (2) alter their visual strategy from viewing the mouth to viewing the eyes of a talker with signal degradations, which negatively affects speech perception.
Collapse
|
34
|
Cañigueral R, Hamilton AFDC. The Role of Eye Gaze During Natural Social Interactions in Typical and Autistic People. Front Psychol 2019; 10:560. [PMID: 30930822 PMCID: PMC6428744 DOI: 10.3389/fpsyg.2019.00560] [Citation(s) in RCA: 58] [Impact Index Per Article: 11.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2018] [Accepted: 02/28/2019] [Indexed: 12/13/2022] Open
Abstract
Social interactions involve complex exchanges of a variety of social signals, such as gaze, facial expressions, speech and gestures. Focusing on the dual function of eye gaze, this review explores how the presence of an audience, communicative purpose and temporal dynamics of gaze allow interacting partners to achieve successful communication. First, we focus on how being watched modulates social cognition and behavior. We then show that the study of interpersonal gaze processing, particularly gaze temporal dynamics, can provide valuable understanding of social behavior in real interactions. We propose that the Interpersonal Gaze Processing model, which combines both sensing and signaling functions of eye gaze, provides a framework to make sense of gaze patterns in live interactions. Finally, we discuss how autistic individuals process the belief in being watched and interpersonal dynamics of gaze, and suggest that systematic manipulation of factors modulating gaze signaling can reveal which aspects of social eye gaze are challenging in autism.
Collapse
Affiliation(s)
- Roser Cañigueral
- Institute of Cognitive Neuroscience, Division of Psychology and Language Sciences, University College London, London, United Kingdom
| | | |
Collapse
|
35
|
Scott H, Batten JP, Kuhn G. Why are you looking at me? It's because I'm talking, but mostly because I'm staring or not doing much. Atten Percept Psychophys 2019; 81:109-118. [PMID: 30353500 PMCID: PMC6315010 DOI: 10.3758/s13414-018-1588-6] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
Abstract
Our attention is particularly driven toward faces, especially the eyes, and there is much debate over the factors that modulate this social attentional orienting. Most of the previous research has presented faces in isolation, and we tried to address this shortcoming by measuring people's eye movements whilst they observe more naturalistic and varied social interactions. Participants' eye movements were monitored whilst they watched three different types of social interactions (monologue, manual activity, active attentional misdirection), which were either accompanied by the corresponding audio as speech or by silence. Our results showed that (1) participants spent more time looking at the face when the person was giving a monologue, than when he/she was carrying out manual activities, and in the latter case they spent more time fixating on the person's hands. (2) Hearing speech significantly increases the amount of time participants spent looking at the face (this effect was relatively small), although this was not accounted for by any increase in mouth-oriented gaze. (3) Participants spent significantly more time fixating on the face when direct eye contact was established, and this drive to establish eye contact was significantly stronger in the manual activities than during the monologue. These results highlight people's strategic top-down control over when they attend to faces and the eyes, and support the view that we use our eyes to signal non-verbal information.
Collapse
Affiliation(s)
- Hannah Scott
- Department of Psychology, Goldsmiths, University of London, New Cross, London, SE14 6NW, UK
| | - Jonathan P Batten
- Department of Psychological Sciences, Birkbeck, University of London, London, UK
| | - Gustav Kuhn
- Department of Psychology, Goldsmiths, University of London, New Cross, London, SE14 6NW, UK.
| |
Collapse
|
36
|
Guérin-Dugué A, Roy RN, Kristensen E, Rivet B, Vercueil L, Tcherkassof A. Temporal Dynamics of Natural Static Emotional Facial Expressions Decoding: A Study Using Event- and Eye Fixation-Related Potentials. Front Psychol 2018; 9:1190. [PMID: 30050487 PMCID: PMC6052106 DOI: 10.3389/fpsyg.2018.01190] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2018] [Accepted: 06/20/2018] [Indexed: 11/28/2022] Open
Abstract
This study aims at examining the precise temporal dynamics of the emotional facial decoding as it unfolds in the brain, according to the emotions displayed. To characterize this processing as it occurs in ecological settings, we focused on unconstrained visual explorations of natural emotional faces (i.e., free eye movements). The General Linear Model (GLM; Smith and Kutas, 2015a,b; Kristensen et al., 2017a) enables such a depiction. It allows deconvolving adjacent overlapping responses of the eye fixation-related potentials (EFRPs) elicited by the subsequent fixations and the event-related potentials (ERPs) elicited at the stimuli onset. Nineteen participants were displayed with spontaneous static facial expressions of emotions (Neutral, Disgust, Surprise, and Happiness) from the DynEmo database (Tcherkassof et al., 2013). Behavioral results on participants' eye movements show that the usual diagnostic features in emotional decoding (eyes for negative facial displays and mouth for positive ones) are consistent with the literature. The impact of emotional category on both the ERPs and the EFRPs elicited by the free exploration of the emotional faces is observed upon the temporal dynamics of the emotional facial expression processing. Regarding the ERP at stimulus onset, there is a significant emotion-dependent modulation of the P2-P3 complex and LPP components' amplitude at the left frontal site for the ERPs computed by averaging. Yet, the GLM reveals the impact of subsequent fixations on the ERPs time-locked on stimulus onset. Results are also in line with the valence hypothesis. The observed differences between the two estimation methods (Average vs. GLM) suggest the predominance of the right hemisphere at the stimulus onset and the implication of the left hemisphere in the processing of the information encoded by subsequent fixations. Concerning the first EFRP, the Lambda response and the P2 component are modulated by the emotion of surprise compared to the neutral emotion, suggesting an impact of high-level factors, in parieto-occipital sites. Moreover, no difference is observed on the second and subsequent EFRP. Taken together, the results stress the significant gain obtained in analyzing the EFRPs using the GLM method and pave the way toward efficient ecological emotional dynamic stimuli analyses.
Collapse
Affiliation(s)
- Anne Guérin-Dugué
- GIPSA-lab, Institute of Engineering, Université Grenoble Alpes, Centre National de la Recherche Scientifique, Grenoble INP, Grenoble, France
| | - Raphaëlle N. Roy
- Department of Conception and Control of Aeronautical and Spatial Vehicles, Institut Supérieur de l’Aéronautique et de l’Espace, Université Fédérale de Toulouse, Toulouse, France
| | - Emmanuelle Kristensen
- GIPSA-lab, Institute of Engineering, Université Grenoble Alpes, Centre National de la Recherche Scientifique, Grenoble INP, Grenoble, France
- Laboratoire InterUniversitaire de Psychologie – Personnalité, Cognition, Changement Social, Université Grenoble Alpes, Université Savoie Mont Blanc, Grenoble, France
| | - Bertrand Rivet
- GIPSA-lab, Institute of Engineering, Université Grenoble Alpes, Centre National de la Recherche Scientifique, Grenoble INP, Grenoble, France
| | - Laurent Vercueil
- Exploration Fonctionnelle du Système Nerveux, Pôle Psychiatrie, Neurologie et Rééducation Neurologique, CHU Grenoble Alpes, Grenoble, France
- Université Grenoble Alpes, Inserm, CHU Grenoble Alpes, Grenoble Institut des Neurosciences, Grenoble, France
| | - Anna Tcherkassof
- Laboratoire InterUniversitaire de Psychologie – Personnalité, Cognition, Changement Social, Université Grenoble Alpes, Université Savoie Mont Blanc, Grenoble, France
| |
Collapse
|
37
|
Król ME. Auditory noise increases the allocation of attention to the mouth, and the eyes pay the price: An eye-tracking study. PLoS One 2018; 13:e0194491. [PMID: 29558514 PMCID: PMC5860771 DOI: 10.1371/journal.pone.0194491] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2017] [Accepted: 03/05/2018] [Indexed: 01/17/2023] Open
Abstract
We investigated the effect of auditory noise added to speech on patterns of looking at faces in 40 toddlers. We hypothesised that noise would increase the difficulty of processing speech, making children allocate more attention to the mouth of the speaker to gain visual speech cues from mouth movements. We also hypothesised that this shift would cause a decrease in fixation time to the eyes, potentially decreasing the ability to monitor gaze. We found that adding noise increased the number of fixations to the mouth area, at the price of a decreased number of fixations to the eyes. Thus, to our knowledge, this is the first study demonstrating a mouth-eyes trade-off between attention allocated to social cues coming from the eyes and linguistic cues coming from the mouth. We also found that children with higher word recognition proficiency and higher average pupil response had an increased likelihood of fixating the mouth, compared to the eyes and the rest of the screen, indicating stronger motivation to decode the speech.
Collapse
Affiliation(s)
- Magdalena Ewa Król
- SWPS University of Social Sciences and Humanities, Faculty in Wrocław, Wrocław, Poland
| |
Collapse
|
38
|
Busin Y, Lukasova K, Asthana MK, Macedo EC. Hemiface Differences in Visual Exploration Patterns When Judging the Authenticity of Facial Expressions. Front Psychol 2018; 8:2332. [PMID: 29367851 PMCID: PMC5767895 DOI: 10.3389/fpsyg.2017.02332] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2017] [Accepted: 12/21/2017] [Indexed: 11/29/2022] Open
Abstract
Past studies have found asymmetry biases in human emotion recognition. The left side bias refers to preferential looking at the left-hemiface when actively exploring face images. However, these studies have been mainly conducted with static and frontally oriented stimuli, whereas real-life emotion recognition takes place on dynamic faces viewed from different angles. The aim of this study was to assess the judgment of genuine vs. masked expressions in dynamic movie clips of faces rotated to the right or left side. Forty-eight participants judged the expressions on faces displaying genuine or masked happy, sad, and fearful emotions. The head of the actor was either rotated to the left by a 45° angle, thus showing the left side of the face (standard orientation), or inverted, with the same face shown from the right side perspective. The eye movements were registered by the eye tracker and the data were analyzed for the inverse efficiency score (IES), the number of fixations, gaze time on the whole face and in the regions of interest. Results showed shorter IESs and gaze times for happy compared to sad and fearful emotions, but no difference was found for these variables between sad and fearful emotions. The left side preference was evident from comparisons of the number of fixations. Standard stimuli received a higher number of fixations than inverted ones. However, gaze time was long on inverted compared to standard faces. Number of fixations on exposed hemiface interacted with the emotions decreasing from happy to sad and fearful. An opposite pattern was found for the occluded hemiface. These results suggest a change in fixation patterns in the rotated faces that may be beneficial for the judgments of expressions. Furthermore, this study replicated the effects of the judgment of genuine and masked emotions using dynamic faces.
Collapse
Affiliation(s)
- Yuri Busin
- Social and Cognitive Neuroscience Laboratory and Developmental Disorders Program, Center for Health and Biological Sciences, Mackenzie Presbyterian University, São Paulo, Brazil
| | - Katerina Lukasova
- Social and Cognitive Neuroscience Laboratory and Developmental Disorders Program, Center for Health and Biological Sciences, Mackenzie Presbyterian University, São Paulo, Brazil.,Center of Mathematics, Computation and Cognition, Federal University of ABC (UFABC), São Bernardo, Brazil
| | - Manish K Asthana
- Department of Humanities and Social Sciences, Indian Institute of Technology Kanpur, Kanpur, India
| | - Elizeu C Macedo
- Social and Cognitive Neuroscience Laboratory and Developmental Disorders Program, Center for Health and Biological Sciences, Mackenzie Presbyterian University, São Paulo, Brazil
| |
Collapse
|
39
|
Looking Behavior and Audiovisual Speech Understanding in Children With Normal Hearing and Children With Mild Bilateral or Unilateral Hearing Loss. Ear Hear 2017; 39:783-794. [PMID: 29252979 DOI: 10.1097/aud.0000000000000534] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES Visual information from talkers facilitates speech intelligibility for listeners when audibility is challenged by environmental noise and hearing loss. Less is known about how listeners actively process and attend to visual information from different talkers in complex multi-talker environments. This study tracked looking behavior in children with normal hearing (NH), mild bilateral hearing loss (MBHL), and unilateral hearing loss (UHL) in a complex multi-talker environment to examine the extent to which children look at talkers and whether looking patterns relate to performance on a speech-understanding task. It was hypothesized that performance would decrease as perceptual complexity increased and that children with hearing loss would perform more poorly than their peers with NH. Children with MBHL or UHL were expected to demonstrate greater attention to individual talkers during multi-talker exchanges, indicating that they were more likely to attempt to use visual information from talkers to assist in speech understanding in adverse acoustics. It also was of interest to examine whether MBHL, versus UHL, would differentially affect performance and looking behavior. DESIGN Eighteen children with NH, eight children with MBHL, and 10 children with UHL participated (8-12 years). They followed audiovisual instructions for placing objects on a mat under three conditions: a single talker providing instructions via a video monitor, four possible talkers alternately providing instructions on separate monitors in front of the listener, and the same four talkers providing both target and nontarget information. Multi-talker background noise was presented at a 5 dB signal-to-noise ratio during testing. An eye tracker monitored looking behavior while children performed the experimental task. RESULTS Behavioral task performance was higher for children with NH than for either group of children with hearing loss. There were no differences in performance between children with UHL and children with MBHL. Eye-tracker analysis revealed that children with NH looked more at the screens overall than did children with MBHL or UHL, though individual differences were greater in the groups with hearing loss. Listeners in all groups spent a small proportion of time looking at relevant screens as talkers spoke. Although looking was distributed across all screens, there was a bias toward the right side of the display. There was no relationship between overall looking behavior and performance on the task. CONCLUSIONS The present study examined the processing of audiovisual speech in the context of a naturalistic task. Results demonstrated that children distributed their looking to a variety of sources during the task, but that children with NH were more likely to look at screens than were those with MBHL/UHL. However, all groups looked at the relevant talkers as they were speaking only a small proportion of the time. Despite variability in looking behavior, listeners were able to follow the audiovisual instructions and children with NH demonstrated better performance than children with MBHL/UHL. These results suggest that performance on some challenging multi-talker audiovisual tasks is not dependent on visual fixation to relevant talkers for children with NH or with MBHL/UHL.
Collapse
|
40
|
Abstract
Eyes have been shown to play a key role during human social interactions. However, to date, no comprehensive cross-discipline model has provided a framework that can account for uniquely human responses to eye cues. In this review, I present a framework that brings together work on the phylogenetic, ontogenetic, and neural bases of perceiving and responding to eyes. Specifically, I argue for a two-process model: a first process that ensures privileged attention to information encoded in the eyes and is important for the detection of other minds and a second process that permits the decoding of information contained in the eyes concerning another person's emotional and mental states. To some degree, these processes are unique to humans, emerge during different times in infant development, can be mapped onto distinct but interconnected brain regions, and likely serve critical functions in facilitating cooperative interactions in humans. I also present evidence to show that oxytocin is a key modulator of sensitive responding to eye cues. Viewing eyes as windows into other minds can therefore be considered a hallmark feature of human social functioning deeply rooted in our biology.
Collapse
|
41
|
De Pascalis L, Kkeli N, Chakrabarti B, Dalton L, Vaillancourt K, Rayson H, Bicknell S, Goodacre T, Cooper P, Stein A, Murray L. Maternal gaze to the infant face: Effects of infant age and facial configuration during mother-infant engagement in the first nine weeks. Infant Behav Dev 2017; 46:91-99. [DOI: 10.1016/j.infbeh.2016.12.003] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2016] [Revised: 12/18/2016] [Accepted: 12/19/2016] [Indexed: 01/21/2023]
|
42
|
Wilson AH, Alsius A, Paré M, Munhall KG. Spatial Frequency Requirements and Gaze Strategy in Visual-Only and Audiovisual Speech Perception. JOURNAL OF SPEECH, LANGUAGE, AND HEARING RESEARCH : JSLHR 2016; 59:601-15. [PMID: 27537379 PMCID: PMC5280058 DOI: 10.1044/2016_jslhr-s-15-0092] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/04/2015] [Revised: 09/16/2015] [Accepted: 10/07/2015] [Indexed: 06/06/2023]
Abstract
PURPOSE The aim of this article is to examine the effects of visual image degradation on performance and gaze behavior in audiovisual and visual-only speech perception tasks. METHOD We presented vowel-consonant-vowel utterances visually filtered at a range of frequencies in visual-only, audiovisual congruent, and audiovisual incongruent conditions (Experiment 1; N = 66). In Experiment 2 (N = 20), participants performed a visual-only speech perception task and in Experiment 3 (N = 20) an audiovisual task while having their gaze behavior monitored using eye-tracking equipment. RESULTS In the visual-only condition, increasing image resolution led to monotonic increases in performance, and proficient speechreaders were more affected by the removal of high spatial information than were poor speechreaders. The McGurk effect also increased with increasing visual resolution, although it was less affected by the removal of high-frequency information. Observers tended to fixate on the mouth more in visual-only perception, but gaze toward the mouth did not correlate with accuracy of silent speechreading or the magnitude of the McGurk effect. CONCLUSIONS The results suggest that individual differences in silent speechreading and the McGurk effect are not related. This conclusion is supported by differential influences of high-resolution visual information on the 2 tasks and differences in the pattern of gaze.
Collapse
Affiliation(s)
- Amanda H. Wilson
- Psychology Department, Queen's University, Kingston, Ontario, Canada
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada
| | - Agnès Alsius
- Psychology Department, Queen's University, Kingston, Ontario, Canada
| | - Martin Paré
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada
| | - Kevin G. Munhall
- Psychology Department, Queen's University, Kingston, Ontario, Canada
- Centre for Neuroscience Studies, Queen's University, Kingston, Ontario, Canada
| |
Collapse
|
43
|
Ayneto A, Sebastian-Galles N. The influence of bilingualism on the preference for the mouth region of dynamic faces. Dev Sci 2016; 20. [PMID: 27196790 DOI: 10.1111/desc.12446] [Citation(s) in RCA: 26] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2015] [Accepted: 03/24/2016] [Indexed: 11/27/2022]
Abstract
Bilingual infants show an extended period of looking at the mouth of talking faces, which provides them with additional articulatory cues that can be used to boost the challenging situation of learning two languages (Pons, Bosch & Lewkowicz, 2015). However, the eye region also provides fundamental cues for emotion perception and recognition, as well as communication. Here, we explored whether the adaptations resulting from learning two languages are specific to linguistic content or if they also influence the focus of attention when looking at dynamic faces. We recorded the eye gaze of bilingual and monolingual infants (8- and 12-month-olds) while watching videos of infants and adults portraying different emotional states (neutral, crying, and laughing). When looking at infant faces, bilinguals looked longer at the mouth region as compared to monolinguals regardless of age. However, when presented with adult faces, 8-month-old bilingual infants looked longer at the mouth region and less at the eye region compared to 8-month-old monolingual infants, but no effect of language exposure was found at 12 months of age. These findings suggest that the bias to the mouth region in bilingual infants at 8 months of age can be generalized to other audiovisual dynamic faces that do not contain linguistic information. We discuss the potential implications of such bias in early social and communicative development.
Collapse
Affiliation(s)
- Alba Ayneto
- Center for Brain and Cognition, Department of Technology, Universitat Pompeu Fabra, Barcelona, Spain
| | - Nuria Sebastian-Galles
- Center for Brain and Cognition, Department of Technology, Universitat Pompeu Fabra, Barcelona, Spain
| |
Collapse
|
44
|
High visual resolution matters in audiovisual speech perception, but only for some. Atten Percept Psychophys 2016; 78:1472-87. [DOI: 10.3758/s13414-016-1109-4] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
|
45
|
Lusk LG, Mitchel AD. Differential Gaze Patterns on Eyes and Mouth During Audiovisual Speech Segmentation. Front Psychol 2016; 7:52. [PMID: 26869959 PMCID: PMC4735377 DOI: 10.3389/fpsyg.2016.00052] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2015] [Accepted: 01/11/2016] [Indexed: 11/17/2022] Open
Abstract
Speech is inextricably multisensory: both auditory and visual components provide critical information for all aspects of speech processing, including speech segmentation, the visual components of which have been the target of a growing number of studies. In particular, a recent study (Mitchel and Weiss, 2014) established that adults can utilize facial cues (i.e., visual prosody) to identify word boundaries in fluent speech. The current study expanded upon these results, using an eye tracker to identify highly attended facial features of the audiovisual display used in Mitchel and Weiss (2014). Subjects spent the most time watching the eyes and mouth. A significant trend in gaze durations was found with the longest gaze duration on the mouth, followed by the eyes and then the nose. In addition, eye-gaze patterns changed across familiarization as subjects learned the word boundaries, showing decreased attention to the mouth in later blocks while attention on other facial features remained consistent. These findings highlight the importance of the visual component of speech processing and suggest that the mouth may play a critical role in visual speech segmentation.
Collapse
Affiliation(s)
- Laina G Lusk
- Neuroscience Program, Bucknell University Lewisburg, PA, USA
| | - Aaron D Mitchel
- Neuroscience Program, Bucknell UniversityLewisburg, PA, USA; Department of Psychology, Bucknell UniversityLewisburg, PA, USA
| |
Collapse
|
46
|
Barenholtz E, Mavica L, Lewkowicz DJ. Language familiarity modulates relative attention to the eyes and mouth of a talker. Cognition 2015; 147:100-5. [PMID: 26649759 DOI: 10.1016/j.cognition.2015.11.013] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2015] [Revised: 10/08/2015] [Accepted: 11/23/2015] [Indexed: 10/22/2022]
Abstract
We investigated whether the audiovisual speech cues available in a talker's mouth elicit greater attention when adults have to process speech in an unfamiliar language vs. a familiar language. Participants performed a speech-encoding task while watching and listening to videos of a talker in a familiar language (English) or an unfamiliar language (Spanish or Icelandic). Attention to the mouth increased in monolingual subjects in response to an unfamiliar language condition but did not in bilingual subjects when the task required speech processing. In the absence of an explicit speech-processing task, subjects attended equally to the eyes and mouth in response to both familiar and unfamiliar languages. Overall, these results demonstrate that language familiarity modulates selective attention to the redundant audiovisual speech cues in a talker's mouth in adults. When our findings are considered together with similar findings from infants, they suggest that this attentional strategy emerges very early in life.
Collapse
Affiliation(s)
- Elan Barenholtz
- Department of Psychology/Center for Complex Systems and Brain Sciences, Florida Atlantic University, Boca Raton, FL, United States.
| | - Lauren Mavica
- Department of Psychology, Florida Atlantic University, Boca Raton, FL, United States
| | - David J Lewkowicz
- Department of Communication Sciences and Disorders, Northeastern University, Boston, MA, United States
| |
Collapse
|
47
|
A link between individual differences in multisensory speech perception and eye movements. Atten Percept Psychophys 2015; 77:1333-41. [PMID: 25810157 DOI: 10.3758/s13414-014-0821-1] [Citation(s) in RCA: 41] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The McGurk effect is an illusion in which visual speech information dramatically alters the perception of auditory speech. However, there is a high degree of individual variability in how frequently the illusion is perceived: some individuals almost always perceive the McGurk effect, while others rarely do. Another axis of individual variability is the pattern of eye movements make while viewing a talking face: some individuals often fixate the mouth of the talker, while others rarely do. Since the talker's mouth carries the visual speech necessary information to induce the McGurk effect, we hypothesized that individuals who frequently perceive the McGurk effect should spend more time fixating the talker's mouth. We used infrared eye tracking to study eye movements as 40 participants viewed audiovisual speech. Frequent perceivers of the McGurk effect were more likely to fixate the mouth of the talker, and there was a significant correlation between McGurk frequency and mouth looking time. The noisy encoding of disparity model of McGurk perception showed that individuals who frequently fixated the mouth had lower sensory noise and higher disparity thresholds than those who rarely fixated the mouth. Differences in eye movements when viewing the talker's face may be an important contributor to interindividual differences in multisensory speech perception.
Collapse
|
48
|
Gavin CJ, Houghton S, Guo K. Dog owners show experience-based viewing behaviour in judging dog face approachability. PSYCHOLOGICAL RESEARCH 2015; 81:75-82. [PMID: 26486649 DOI: 10.1007/s00426-015-0718-1] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2015] [Accepted: 10/09/2015] [Indexed: 11/29/2022]
Abstract
Our prior visual experience plays a critical role in face perception. We show superior perceptual performance for differentiating conspecific (vs non-conspecific), own-race (vs other-race) and familiar (vs unfamiliar) faces. However, it remains unclear whether our experience with faces of other species would influence our gaze allocation for extracting salient facial information. In this eye-tracking study, we asked both dog owners and non-owners to judge the approachability of human, monkey and dog faces, and systematically compared their behavioural performance and gaze pattern associated with the task. Compared to non-owners, dog owners assessed dog faces with shorter time and fewer fixations, but gave higher approachability ratings. The gaze allocation within local facial features was also modulated by the ownership. The averaged proportion of the fixations and viewing time directed at the dog mouth region were significantly less for the dog owners, and more experienced dog owners tended to look more at the dog eyes, suggesting the adoption of a prior experience-based viewing behaviour for assessing dog approachability. No differences in behavioural performance and gaze pattern were observed between dog owners and non-owners when judging human and monkey faces, implying that the dog owner's experience-based gaze strategy for viewing dog faces was not transferable across faces of other species.
Collapse
Affiliation(s)
- Carla Jade Gavin
- School of Psychology, University of Lincoln, Brayford Pool, Lincoln, LN6 7TS, UK
| | - Sarah Houghton
- School of Psychology, University of Lincoln, Brayford Pool, Lincoln, LN6 7TS, UK
| | - Kun Guo
- School of Psychology, University of Lincoln, Brayford Pool, Lincoln, LN6 7TS, UK.
| |
Collapse
|
49
|
Grossman RB, Steinhart E, Mitchell T, McIlvane W. "Look who's talking!" Gaze Patterns for Implicit and Explicit Audio-Visual Speech Synchrony Detection in Children With High-Functioning Autism. Autism Res 2015; 8:307-16. [PMID: 25620208 DOI: 10.1002/aur.1447] [Citation(s) in RCA: 34] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2013] [Accepted: 11/25/2014] [Indexed: 11/11/2022]
Abstract
Conversation requires integration of information from faces and voices to fully understand the speaker's message. To detect auditory-visual asynchrony of speech, listeners must integrate visual movements of the face, particularly the mouth, with auditory speech information. Individuals with autism spectrum disorder may be less successful at such multisensory integration, despite their demonstrated preference for looking at the mouth region of a speaker. We showed participants (individuals with and without high-functioning autism (HFA) aged 8-19) a split-screen video of two identical individuals speaking side by side. Only one of the speakers was in synchrony with the corresponding audio track and synchrony switched between the two speakers every few seconds. Participants were asked to watch the video without further instructions (implicit condition) or to specifically watch the in-synch speaker (explicit condition). We recorded which part of the screen and face their eyes targeted. Both groups looked at the in-synch video significantly more with explicit instructions. However, participants with HFA looked at the in-synch video less than typically developing (TD) peers and did not increase their gaze time as much as TD participants in the explicit task. Importantly, the HFA group looked significantly less at the mouth than their TD peers, and significantly more at non-face regions of the image. There were no between-group differences for eye-directed gaze. Overall, individuals with HFA spend less time looking at the crucially important mouth region of the face during auditory-visual speech integration, which is maladaptive gaze behavior for this type of task.
Collapse
Affiliation(s)
- Ruth B Grossman
- Emerson College, Department of Communication Sciences and Disorders, 120 Boylston Street, Boston, Massachusetts.,University of Massachusetts Medical School Shriver Center, 200 Trapelo Rd, Waltham, Massachusetts
| | - Erin Steinhart
- University of Massachusetts Medical School Shriver Center, 200 Trapelo Rd, Waltham, Massachusetts
| | - Teresa Mitchell
- University of Massachusetts Medical School Shriver Center, 200 Trapelo Rd, Waltham, Massachusetts
| | - William McIlvane
- University of Massachusetts Medical School Shriver Center, 200 Trapelo Rd, Waltham, Massachusetts
| |
Collapse
|
50
|
Rhodes T, Kello CT, Kerster B. Intrinsic and extrinsic contributions to heavy tails in visual foraging. VISUAL COGNITION 2014. [DOI: 10.1080/13506285.2014.918070] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
|