1
|
Li S, Zhang T, Yang F, Li X, Wang Z, Zhao D. A Dynamic Multi-Scale Convolution Model for Face Recognition Using Event-Related Potentials. SENSORS (BASEL, SWITZERLAND) 2024; 24:4368. [PMID: 39001147 PMCID: PMC11244416 DOI: 10.3390/s24134368] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 06/21/2024] [Accepted: 06/27/2024] [Indexed: 07/16/2024]
Abstract
With the development of data mining technology, the analysis of event-related potential (ERP) data has evolved from statistical analysis of time-domain features to data-driven techniques based on supervised and unsupervised learning. However, there are still many challenges in understanding the relationship between ERP components and the representation of familiar and unfamiliar faces. To address this, this paper proposes a model based on Dynamic Multi-Scale Convolution for group recognition of familiar and unfamiliar faces. This approach uses generated weight masks for cross-subject familiar/unfamiliar face recognition using a multi-scale model. The model employs a variable-length filter generator to dynamically determine the optimal filter length for time-series samples, thereby capturing features at different time scales. Comparative experiments are conducted to evaluate the model's performance against SOTA models. The results demonstrate that our model achieves impressive outcomes, with a balanced accuracy rate of 93.20% and an F1 score of 88.54%, outperforming the methods used for comparison. The ERP data extracted from different time regions in the model can also provide data-driven technical support for research based on the representation of different ERP components.
Collapse
Affiliation(s)
- Shengkai Li
- School of Automation, Qingdao University, Qingdao 266071, China
- State Key Laboratory of Multimodal Artifcial Intelligence Systems, The Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| | - Tonglin Zhang
- State Key Laboratory of Multimodal Artifcial Intelligence Systems, The Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100190, China
| | - Fangmei Yang
- State Key Laboratory of Multimodal Artifcial Intelligence Systems, The Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| | - Xian Li
- School of Automation, Qingdao University, Qingdao 266071, China
- Shandong Key Laboratory of Industrial Control Technology, Qingdao University, Qingdao 266071, China
| | - Ziyang Wang
- State Key Laboratory of Multimodal Artifcial Intelligence Systems, The Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| | - Dongjie Zhao
- School of Automation, Qingdao University, Qingdao 266071, China
- Shandong Key Laboratory of Industrial Control Technology, Qingdao University, Qingdao 266071, China
| |
Collapse
|
2
|
Kopal J, Hlinka J, Despouy E, Valton L, Denuelle M, Sol J, Curot J, Barbeau EJ. Large-scale network dynamics underlying the first few hundred milliseconds after stimulus presentation: An investigation of visual recognition memory using iEEG. Hum Brain Mapp 2023; 44:5795-5809. [PMID: 37688546 PMCID: PMC10619408 DOI: 10.1002/hbm.26477] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2023] [Revised: 08/14/2023] [Accepted: 08/23/2023] [Indexed: 09/11/2023] Open
Abstract
Recognition memory is the ability to recognize previously encountered objects. Even this relatively simple, yet extremely fast, ability requires the coordinated activity of large-scale brain networks. However, little is known about the sub-second dynamics of these networks. The majority of current studies into large-scale network dynamics is primarily based on imaging techniques suffering from either poor temporal or spatial resolution. We investigated the dynamics of large-scale functional brain networks underlying recognition memory at the millisecond scale. Specifically, we analyzed dynamic effective connectivity from intracranial electroencephalography while epileptic subjects (n = 18) performed a fast visual recognition memory task. Our data-driven investigation using Granger causality and the analysis of communities with the Louvain algorithm spotlighted a dynamic interplay of two large-scale networks associated with successful recognition. The first network involved the right visual ventral stream and bilateral frontal regions. It was characterized by early, predominantly bottom-up information flow peaking at 115 ms. It was followed by the involvement of another network with predominantly top-down connectivity peaking at 220 ms, mainly in the left anterior hemisphere. The transition between these two networks was associated with changes in network topology, evolving from a more segregated to a more integrated state. These results highlight that distinct large-scale brain networks involved in visual recognition memory unfold early and quickly, within the first 300 ms after stimulus onset. Our study extends the current understanding of the rapid network changes during rapid cognitive processes.
Collapse
Affiliation(s)
- Jakub Kopal
- Institute of Computer Science of the Czech Academy of SciencesPragueCzech Republic
- Department of Computing and Control EngineeringUniversity of Chemistry and TechnologyPragueCzech Republic
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
| | - Jaroslav Hlinka
- Institute of Computer Science of the Czech Academy of SciencesPragueCzech Republic
- National Institute of Mental HealthKlecanyCzech Republic
| | - Elodie Despouy
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
| | - Luc Valton
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
- University Hospital PurpanToulouseFrance
| | - Marie Denuelle
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
- University Hospital PurpanToulouseFrance
| | - Jean‐Christophe Sol
- University Hospital PurpanToulouseFrance
- Toulouse NeuroImaging CenterToulouseFrance
| | - Jonathan Curot
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
- University Hospital PurpanToulouseFrance
| | - Emmanuel J. Barbeau
- Centre de Recherche Cerveau et CognitionToulouse III University – CNRS UMR 5549ToulouseFrance
| |
Collapse
|
3
|
Bruchmann M, Mertens L, Schindler S, Straube T. Potentiated early neural responses to fearful faces are not driven by specific face parts. Sci Rep 2023; 13:4613. [PMID: 36944705 PMCID: PMC10030637 DOI: 10.1038/s41598-023-31752-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Accepted: 03/16/2023] [Indexed: 03/23/2023] Open
Abstract
Prioritized processing of fearful compared to neutral faces is reflected in increased amplitudes of components of the event-related potential (ERP). It is unknown whether specific face parts drive these modulations. Here, we investigated the contributions of face parts on ERPs to task-irrelevant fearful and neutral faces using an ERP-dependent facial decoding technique and a large sample of participants (N = 83). Classical ERP analyses showed typical and robust increases of N170 and EPN amplitudes by fearful relative to neutral faces. Facial decoding further showed that the absolute amplitude of these components, as well as the P1, was driven by the low-frequency contrast of specific face parts. However, the difference between fearful and neutral faces was not driven by any specific face part, as supported by Bayesian statistics. Furthermore, there were no correlations between trait anxiety and main effects or interactions. These results suggest that increased N170 and EPN amplitudes to task-irrelevant fearful compared to neutral faces are not driven by specific facial regions but represent a holistic face processing effect.
Collapse
Affiliation(s)
- Maximilian Bruchmann
- Institute of Medical Psychology and Systems Neuroscience, University of Muenster, Von-Esmarch-Str. 52, 48149, Münster, Germany.
- Otto Creutzfeldt Center for Cognitive and Behavioral Neuroscience, University of Muenster, Münster, Germany.
| | - Léa Mertens
- Institute of Medical Psychology and Systems Neuroscience, University of Muenster, Von-Esmarch-Str. 52, 48149, Münster, Germany
| | - Sebastian Schindler
- Institute of Medical Psychology and Systems Neuroscience, University of Muenster, Von-Esmarch-Str. 52, 48149, Münster, Germany
- Otto Creutzfeldt Center for Cognitive and Behavioral Neuroscience, University of Muenster, Münster, Germany
| | - Thomas Straube
- Institute of Medical Psychology and Systems Neuroscience, University of Muenster, Von-Esmarch-Str. 52, 48149, Münster, Germany
- Otto Creutzfeldt Center for Cognitive and Behavioral Neuroscience, University of Muenster, Münster, Germany
| |
Collapse
|
4
|
Hu Z, Zhang Z, Liang Z, Zhang L, Li L, Huang G. A New Perspective on Individual Reliability beyond Group Effect for Event-related Potentials: A Multisensory Investigation and Computational Modeling. Neuroimage 2022; 250:118937. [PMID: 35091080 DOI: 10.1016/j.neuroimage.2022.118937] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Revised: 01/14/2022] [Accepted: 01/24/2022] [Indexed: 10/19/2022] Open
Abstract
The dominant approach in investigating the individual reliability for event-related potentials (ERPs) is to extract peak-related features at electrodes showing the strongest group effects. Such a peak-based approach implicitly assumes ERP components showing a stronger group effect are also more reliable, but this assumption has not been substantially validated and few studies have investigated the reliability of ERPs beyond peaks. In this study, we performed a rigorous evaluation of the test-retest reliability of ERPs collected in a multisensory and cognitive experiment from 82 healthy adolescents, each having two sessions. By comparing group effects and individual reliability, we found that a stronger group-level response in ERPs did not guarantee higher reliability. A perspective of neural oscillation should be adopted for the analysis of reliability. Further, by simulating ERPs with an oscillation-based computational model, we found that the consistency between group-level ERP responses and individual reliability was modulated by inter-subject latency jitter and inter-trial variability. The current findings suggest that the conventional peak-based approach may underestimate the individual reliability in ERPs and a neural oscillation perspective on ERP reliability should be considered. Hence, a comprehensive evaluation of the reliability of ERP measurements should be considered in individual-level neurophysiological trait evaluation and psychiatric disorder diagnosis.
Collapse
Affiliation(s)
- Zhenxing Hu
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China
| | - Zhiguo Zhang
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China; Peng Cheng Laboratory, Shenzhen, Guangdong, 518055, China; Marshall Laboratory of Biomedical Engineering, Shenzhen University, Shenzhen, 518060, China
| | - Zhen Liang
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China
| | - Li Zhang
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China
| | - Linling Li
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China
| | - Gan Huang
- School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518060, China; Guangdong Provincial Key Laboratory of Biomedical Measurements and Ultrasound Imaging, Shenzhen University, Shenzhen, Guangdong, 518060, China.
| |
Collapse
|
5
|
Liu M, Duan Y, Ince RAA, Chen C, Garrod OGB, Schyns PG, Jack RE. Facial expressions elicit multiplexed perceptions of emotion categories and dimensions. Curr Biol 2022; 32:200-209.e6. [PMID: 34767768 PMCID: PMC8751635 DOI: 10.1016/j.cub.2021.10.035] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2021] [Revised: 09/07/2021] [Accepted: 10/14/2021] [Indexed: 11/22/2022]
Abstract
Human facial expressions are complex, multi-component signals that can communicate rich information about emotions,1-5 including specific categories, such as "anger," and broader dimensions, such as "negative valence, high arousal."6-8 An enduring question is how this complex signaling is achieved. Communication theory predicts that multi-component signals could transmit each type of emotion information-i.e., specific categories and broader dimensions-via the same or different facial signal components, with implications for elucidating the system and ontology of facial expression communication.9 We addressed this question using a communication-systems-based method that agnostically generates facial expressions and uses the receiver's perceptions to model the specific facial signal components that represent emotion category and dimensional information to them.10-12 First, we derived the facial expressions that elicit the perception of emotion categories (i.e., the six classic emotions13 plus 19 complex emotions3) and dimensions (i.e., valence and arousal) separately, in 60 individual participants. Comparison of these facial signals showed that they share subsets of components, suggesting that specific latent signals jointly represent-i.e., multiplex-categorical and dimensional information. Further examination revealed these specific latent signals and the joint information they represent. Our results-based on white Western participants, same-ethnicity face stimuli, and commonly used English emotion terms-show that facial expressions can jointly represent specific emotion categories and broad dimensions to perceivers via multiplexed facial signal components. Our results provide insights into the ontology and system of facial expression communication and a new information-theoretic framework that can characterize its complexities.
Collapse
Affiliation(s)
- Meng Liu
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Yaocong Duan
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Robin A A Ince
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Chaona Chen
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Oliver G B Garrod
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Philippe G Schyns
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK
| | - Rachael E Jack
- School of Psychology & Neuroscience, University of Glasgow, 62 Hillhead Street, Glasgow G12 8QB, UK.
| |
Collapse
|
6
|
Marcolin F, Vezzetti E, Monaci M. Face perception foundations for pattern recognition algorithms. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.02.074] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
7
|
Jaworska K, Yi F, Ince RAA, van Rijsbergen NJ, Schyns PG, Rousselet GA. Healthy aging delays the neural processing of face features relevant for behavior by 40 ms. Hum Brain Mapp 2019; 41:1212-1225. [PMID: 31782861 PMCID: PMC7268067 DOI: 10.1002/hbm.24869] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2019] [Revised: 10/16/2019] [Accepted: 11/10/2019] [Indexed: 12/18/2022] Open
Abstract
Fast and accurate face processing is critical for everyday social interactions, but it declines and becomes delayed with age, as measured by both neural and behavioral responses. Here, we addressed the critical challenge of understanding how aging changes neural information processing mechanisms to delay behavior. Young (20-36 years) and older (60-86 years) adults performed the basic social interaction task of detecting a face versus noise while we recorded their electroencephalogram (EEG). In each participant, using a new information theoretic framework we reconstructed the features supporting face detection behavior, and also where, when and how EEG activity represents them. We found that occipital-temporal pathway activity dynamically represents the eyes of the face images for behavior ~170 ms poststimulus, with a 40 ms delay in older adults that underlies their 200 ms behavioral deficit of slower reaction times. Our results therefore demonstrate how aging can change neural information processing mechanisms that underlie behavioral slow down.
Collapse
Affiliation(s)
- Katarzyna Jaworska
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, UK
| | - Fei Yi
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, UK
| | - Robin A A Ince
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, UK
| | | | - Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, UK
| | | |
Collapse
|
8
|
Zhan J, Ince RAA, van Rijsbergen N, Schyns PG. Dynamic Construction of Reduced Representations in the Brain for Perceptual Decision Behavior. Curr Biol 2019; 29:319-326.e4. [PMID: 30639108 PMCID: PMC6345582 DOI: 10.1016/j.cub.2018.11.049] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2018] [Revised: 10/23/2018] [Accepted: 11/20/2018] [Indexed: 01/03/2023]
Abstract
Over the past decade, extensive studies of the brain regions that support face, object, and scene recognition suggest that these regions have a hierarchically organized architecture that spans the occipital and temporal lobes [1-14], where visual categorizations unfold over the first 250 ms of processing [15-19]. This same architecture is flexibly involved in multiple tasks that require task-specific representations-e.g. categorizing the same object as "a car" or "a Porsche." While we partly understand where and when these categorizations happen in the occipito-ventral pathway, the next challenge is to unravel how these categorizations happen. That is, how does high-dimensional input collapse in the occipito-ventral pathway to become low dimensional representations that guide behavior? To address this, we investigated what information the brain processes in a visual perception task and visualized the dynamic representation of this information in brain activity. To do so, we developed stimulus information representation (SIR), an information theoretic framework, to tease apart stimulus information that supports behavior from that which does not. We then tracked the dynamic representations of both in magneto-encephalographic (MEG) activity. Using SIR, we demonstrate that a rapid (∼170 ms) reduction of behaviorally irrelevant information occurs in the occipital cortex and that representations of the information that supports distinct behaviors are constructed in the right fusiform gyrus (rFG). Our results thus highlight how SIR can be used to investigate the component processes of the brain by considering interactions between three variables (stimulus information, brain activity, behavior), rather than just two, as is the current norm.
Collapse
Affiliation(s)
- Jiayu Zhan
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, United Kingdom
| | - Robin A A Ince
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, United Kingdom
| | - Nicola van Rijsbergen
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, United Kingdom
| | - Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, United Kingdom; School of Psychology, University of Glasgow, 62 Hillhead Street, Glasgow, Scotland G12 8QB, United Kingdom.
| |
Collapse
|
9
|
Estéphan A, Fiset D, Saumure C, Plouffe-Demers MP, Zhang Y, Sun D, Blais C. Time Course of Cultural Differences in Spatial Frequency Use for Face Identification. Sci Rep 2018; 8:1816. [PMID: 29379032 PMCID: PMC5788938 DOI: 10.1038/s41598-018-19971-1] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2017] [Accepted: 12/15/2017] [Indexed: 11/13/2022] Open
Abstract
Several previous studies of eye movements have put forward that, during face recognition, Easterners spread their attention across a greater part of their visual field than Westerners. Recently, we found that culture’s effect on the perception of faces reaches mechanisms deeper than eye movements, therefore affecting the very nature of information sampled by the visual system: that is, Westerners globally rely more than Easterners on fine-grained visual information (i.e. high spatial frequencies; SFs), whereas Easterners rely more on coarse-grained visual information (i.e. low SFs). These findings suggest that culture influences basic visual processes; however, the temporal onset and dynamics of these culture-specific perceptual differences are still unknown. Here, we investigate the time course of SF use in Western Caucasian (Canadian) and East Asian (Chinese) observers during a face identification task. Firstly, our results confirm that Easterners use relatively lower SFs than Westerners, while the latter use relatively higher SFs. More importantly, our results indicate that these differences arise as early as 34 ms after stimulus onset, and remain stable through time. Our research supports the hypothesis that Westerners and Easterners initially rely on different types of visual information during face processing.
Collapse
Affiliation(s)
- Amanda Estéphan
- Département de psychoéducation et de psychologie, Université du Québec en Outaouais, Québec, Canada.,Département de psychologie, Université du Québec à Montréal, Québec, Canada
| | - Daniel Fiset
- Département de psychoéducation et de psychologie, Université du Québec en Outaouais, Québec, Canada
| | - Camille Saumure
- Département de psychoéducation et de psychologie, Université du Québec en Outaouais, Québec, Canada
| | | | - Ye Zhang
- Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China.,Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
| | - Dan Sun
- Institute of Psychological Science, Hangzhou Normal University, Hangzhou, China.,Zhejiang Key Laboratory for Research in Assessment of Cognitive Impairments, Hangzhou, China
| | - Caroline Blais
- Département de psychoéducation et de psychologie, Université du Québec en Outaouais, Québec, Canada.
| |
Collapse
|
10
|
Affiliation(s)
- Rachael E. Jack
- Institute of Neuroscience and Psychology, and School of Psychology, University of Glasgow, Glasgow G12 8QB United Kingdom;
| | - Philippe G. Schyns
- Institute of Neuroscience and Psychology, and School of Psychology, University of Glasgow, Glasgow G12 8QB United Kingdom;
| |
Collapse
|
11
|
A cognitive architecture account of the visual local advantage phenomenon in autism spectrum disorders. Vision Res 2016; 126:278-290. [DOI: 10.1016/j.visres.2015.04.009] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2014] [Revised: 02/17/2015] [Accepted: 04/14/2015] [Indexed: 11/24/2022]
|
12
|
Ince RAA, Jaworska K, Gross J, Panzeri S, van Rijsbergen NJ, Rousselet GA, Schyns PG. The Deceptively Simple N170 Reflects Network Information Processing Mechanisms Involving Visual Feature Coding and Transfer Across Hemispheres. Cereb Cortex 2016; 26:4123-4135. [PMID: 27550865 PMCID: PMC5066825 DOI: 10.1093/cercor/bhw196] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
A key to understanding visual cognition is to determine “where”, “when”, and “how” brain responses reflect the processing of the specific visual features that modulate categorization behavior—the “what”. The N170 is the earliest Event-Related Potential (ERP) that preferentially responds to faces. Here, we demonstrate that a paradigmatic shift is necessary to interpret the N170 as the product of an information processing network that dynamically codes and transfers face features across hemispheres, rather than as a local stimulus-driven event. Reverse-correlation methods coupled with information-theoretic analyses revealed that visibility of the eyes influences face detection behavior. The N170 initially reflects coding of the behaviorally relevant eye contralateral to the sensor, followed by a causal communication of the other eye from the other hemisphere. These findings demonstrate that the deceptively simple N170 ERP hides a complex network information processing mechanism involving initial coding and subsequent cross-hemispheric transfer of visual features.
Collapse
Affiliation(s)
- Robin A A Ince
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8QB, UK
| | - Katarzyna Jaworska
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8QB, UK
| | - Joachim Gross
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8QB, UK
| | - Stefano Panzeri
- Laboratory of Neural Computation, Istituto Italiano di Tecnologia, Rovereto 38068, Italy
| | | | - Guillaume A Rousselet
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8QB, UK
| | - Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8QB, UK
| |
Collapse
|
13
|
Abstract
As a highly social species, humans frequently exchange social information to support almost all facets of life. One of the richest and most powerful tools in social communication is the face, from which observers can quickly and easily make a number of inferences - about identity, gender, sex, age, race, ethnicity, sexual orientation, physical health, attractiveness, emotional state, personality traits, pain or physical pleasure, deception, and even social status. With the advent of the digital economy, increasing globalization and cultural integration, understanding precisely which face information supports social communication and which produces misunderstanding is central to the evolving needs of modern society (for example, in the design of socially interactive digital avatars and companion robots). Doing so is challenging, however, because the face can be thought of as comprising a high-dimensional, dynamic information space, and this impacts cognitive science and neuroimaging, and their broader applications in the digital economy. New opportunities to address this challenge are arising from the development of new methods and technologies, coupled with the emergence of a modern scientific culture that embraces cross-disciplinary approaches. Here, we briefly review one such approach that combines state-of-the-art computer graphics, psychophysics and vision science, cultural psychology and social cognition, and highlight the main knowledge advances it has generated. In the light of current developments, we provide a vision of the future directions in the field of human facial communication within and across cultures.
Collapse
Affiliation(s)
- Rachael E Jack
- School of Psychology, University of Glasgow, Scotland G12 8QB, UK; Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, UK.
| | - Philippe G Schyns
- School of Psychology, University of Glasgow, Scotland G12 8QB, UK; Institute of Neuroscience and Psychology, University of Glasgow, Scotland G12 8QB, UK.
| |
Collapse
|
14
|
Almeida PR, Ferreira-Santos F, Chaves PL, Paiva TO, Barbosa F, Marques-Teixeira J. Perceived arousal of facial expressions of emotion modulates the N170, regardless of emotional category: Time domain and time–frequency dynamics. Int J Psychophysiol 2016; 99:48-56. [DOI: 10.1016/j.ijpsycho.2015.11.017] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2015] [Revised: 10/13/2015] [Accepted: 11/30/2015] [Indexed: 10/22/2022]
|
15
|
Meinhardt G, Meinhardt-Injac B, Persike M. The complete design in the composite face paradigm: role of response bias, target certainty, and feedback. Front Hum Neurosci 2014; 8:885. [PMID: 25400573 PMCID: PMC4215786 DOI: 10.3389/fnhum.2014.00885] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2014] [Accepted: 10/14/2014] [Indexed: 11/25/2022] Open
Abstract
Some years ago an improved design (the “complete design”) was proposed to assess the composite face effect in terms of a congruency effect, defined as the performance difference for congruent and incongruent target to no-target relationships (Cheung et al., 2008). In a recent paper Rossion (2013) questioned whether the congruency effect was a valid hallmark of perceptual integration, because it may contain confounds with face-unspecific interference effects. Here we argue that the complete design is well-balanced and allows one to separate face-specific from face-unspecific effects. We used the complete design for a same/different composite stimulus matching task with face and non-face objects (watches). Subjects performed the task with and without trial-by-trial feedback, and with low and high certainty about the target half. Results showed large congruency effects for faces, particularly when subjects were informed late in the trial about which face halves had to be matched. Analysis of response bias revealed that subjects preferred the “different” response in incongruent trials, which is expected when upper and lower face halves are integrated perceptually at the encoding stage. The results pattern was observed in the absence of feedback, while providing feedback generally attenuated the congruency effect, and led to an avoidance of response bias. For watches no or marginal congruency effects and a moderate global “same” bias were observed. We conclude that the congruency effect, when complemented by an evaluation of response bias, is a valid hallmark of feature integration that allows one to separate faces from non-face objects.
Collapse
Affiliation(s)
- Günter Meinhardt
- Department of Psychology, Johannes Gutenberg University Mainz Mainz, Germany
| | | | - Malte Persike
- Department of Psychology, Johannes Gutenberg University Mainz Mainz, Germany
| |
Collapse
|
16
|
Rossion B. Understanding face perception by means of human electrophysiology. Trends Cogn Sci 2014; 18:310-8. [DOI: 10.1016/j.tics.2014.02.013] [Citation(s) in RCA: 153] [Impact Index Per Article: 15.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2014] [Revised: 02/26/2014] [Accepted: 02/27/2014] [Indexed: 10/25/2022]
|
17
|
Abstract
Previous magnetoencephalography/electroencephalography (M/EEG) studies have suggested that face processing is extremely rapid, indeed faster than any other object category. Most studies, however, have been performed using centered, cropped stimuli presented on a blank background resulting in artificially low interstimulus variability. In contrast, the aim of the present study was to assess the underlying temporal dynamics of face detection presented in complex natural scenes. We recorded EEG activity while participants performed a rapid go/no-go categorization task in which they had to detect the presence of a human face. Subjects performed at ceiling (94.8% accuracy), and traditional event-related potential analyses revealed only modest modulations of the two main components classically associated with face processing (P100 and N170). A multivariate pattern analysis conducted across all EEG channels revealed that face category could, however, be readout very early, under 100 ms poststimulus onset. Decoding was linked to reaction time as early as 125 ms. Decoding accuracy did not increase monotonically; we report an increase during an initial 95-140 ms period followed by a plateau ∼140-185 ms-perhaps reflecting a transitory stabilization of the face information available-and a strong increase afterward. Further analyses conducted on individual images confirmed these phases, further suggesting that decoding accuracy may be initially driven by low-level stimulus properties. Such latencies appear to be surprisingly short given the complexity of the natural scenes and the large intraclass variability of the face stimuli used, suggesting that the visual system is highly optimized for the processing of natural scenes.
Collapse
|
18
|
Liu J, Li J, Feng L, Li L, Tian J, Lee K. Seeing Jesus in toast: neural and behavioral correlates of face pareidolia. Cortex 2014; 53:60-77. [PMID: 24583223 DOI: 10.1016/j.cortex.2014.01.013] [Citation(s) in RCA: 98] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2013] [Revised: 11/05/2013] [Accepted: 01/21/2014] [Indexed: 10/25/2022]
Abstract
Face pareidolia is the illusory perception of non-existent faces. The present study, for the first time, contrasted behavioral and neural responses of face pareidolia with those of letter pareidolia to explore face-specific behavioral and neural responses during illusory face processing. Participants were shown pure-noise images but were led to believe that 50% of them contained either faces or letters; they reported seeing faces or letters illusorily 34% and 38% of the time, respectively. The right fusiform face area (rFFA) showed a specific response when participants "saw" faces as opposed to letters in the pure-noise images. Behavioral responses during face pareidolia produced a classification image (CI) that resembled a face, whereas those during letter pareidolia produced a CI that was letter-like. Further, the extent to which such behavioral CIs resembled faces was directly related to the level of face-specific activations in the rFFA. This finding suggests that the rFFA plays a specific role not only in processing of real faces but also in illusory face perception, perhaps serving to facilitate the interaction between bottom-up information from the primary visual cortex and top-down signals from the prefrontal cortex (PFC). Whole brain analyses revealed a network specialized in face pareidolia, including both the frontal and occipitotemporal regions. Our findings suggest that human face processing has a strong top-down component whereby sensory input with even the slightest suggestion of a face can result in the interpretation of a face.
Collapse
Affiliation(s)
- Jiangang Liu
- School of Computer and Information Technology, Beijing Jiaotong University, Beijing, China; Dr. Eric Jackman Institute of Child Study, University of Toronto, Toronto, Canada
| | - Jun Li
- School of Life Science and Technology, Xidian University, Xi'an, China
| | - Lu Feng
- Institute of Automation Chinese Academy of Sciences, Beijing, China
| | - Ling Li
- School of Computer and Information Technology, Beijing Jiaotong University, Beijing, China
| | - Jie Tian
- School of Life Science and Technology, Xidian University, Xi'an, China; Institute of Automation Chinese Academy of Sciences, Beijing, China.
| | - Kang Lee
- Dr. Eric Jackman Institute of Child Study, University of Toronto, Toronto, Canada.
| |
Collapse
|
19
|
Jack RE, Garrod OGB, Schyns PG. Dynamic facial expressions of emotion transmit an evolving hierarchy of signals over time. Curr Biol 2014; 24:187-192. [PMID: 24388852 DOI: 10.1016/j.cub.2013.11.064] [Citation(s) in RCA: 191] [Impact Index Per Article: 19.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2013] [Revised: 11/28/2013] [Accepted: 11/29/2013] [Indexed: 10/25/2022]
Abstract
Designed by biological and social evolutionary pressures, facial expressions of emotion comprise specific facial movements to support a near-optimal system of signaling and decoding. Although highly dynamical, little is known about the form and function of facial expression temporal dynamics. Do facial expressions transmit diagnostic signals simultaneously to optimize categorization of the six classic emotions, or sequentially to support a more complex communication system of successive categorizations over time? Our data support the latter. Using a combination of perceptual expectation modeling, information theory, and Bayesian classifiers, we show that dynamic facial expressions of emotion transmit an evolving hierarchy of "biologically basic to socially specific" information over time. Early in the signaling dynamics, facial expressions systematically transmit few, biologically rooted face signals supporting the categorization of fewer elementary categories (e.g., approach/avoidance). Later transmissions comprise more complex signals that support categorization of a larger number of socially specific categories (i.e., the six classic emotions). Here, we show that dynamic facial expressions of emotion provide a sophisticated signaling system, questioning the widely accepted notion that emotion communication is comprised of six basic (i.e., psychologically irreducible) categories, and instead suggesting four.
Collapse
Affiliation(s)
- Rachael E Jack
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, Scotland, G12 8QB, UK.
| | - Oliver G B Garrod
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, Scotland, G12 8QB, UK
| | - Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, Scotland, G12 8QB, UK
| |
Collapse
|
20
|
|
21
|
Neath KN, Itier RJ. Facial expression discrimination varies with presentation time but not with fixation on features: a backward masking study using eye-tracking. Cogn Emot 2013; 28:115-31. [PMID: 23879672 DOI: 10.1080/02699931.2013.812557] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
The current study investigated the effects of presentation time and fixation to expression-specific diagnostic features on emotion discrimination performance, in a backward masking task. While no differences were found when stimuli were presented for 16.67 ms, differences between facial emotions emerged beyond the happy-superiority effect at presentation times as early as 50 ms. Happy expressions were best discriminated, followed by neutral and disgusted, then surprised, and finally fearful expressions presented for 50 and 100 ms. While performance was not improved by the use of expression-specific diagnostic facial features, performance increased with presentation time for all emotions. Results support the idea of an integration of facial features (holistic processing) varying as a function of emotion and presentation time.
Collapse
Affiliation(s)
- Karly N Neath
- a Department of Psychology , University of Waterloo , Waterloo , ON , Canada
| | | |
Collapse
|
22
|
Du S, Martinez AM. Wait, are you sad or angry? Large exposure time differences required for the categorization of facial expressions of emotion. J Vis 2013; 13:13. [PMID: 23509409 DOI: 10.1167/13.4.13] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Facial expressions of emotion are essential components of human behavior, yet little is known about the hierarchical organization of their cognitive analysis. We study the minimum exposure time needed to successfully classify the six classical facial expressions of emotion (joy, surprise, sadness, anger, disgust, fear) plus neutral as seen at different image resolutions (240 × 160 to 15 × 10 pixels). Our results suggest a consistent hierarchical analysis of these facial expressions regardless of the resolution of the stimuli. Happiness and surprise can be recognized after very short exposure times (10-20 ms), even at low resolutions. Fear and anger are recognized the slowest (100-250 ms), even in high-resolution images, suggesting a later computation. Sadness and disgust are recognized in between (70-200 ms). The minimum exposure time required for successful classification of each facial expression correlates with the ability of a human subject to identify it correctly at low resolutions. These results suggest a fast, early computation of expressions represented mostly by low spatial frequencies or global configural cues and a later, slower process for those categories requiring a more fine-grained analysis of the image. We also demonstrate that those expressions that are mostly visible in higher-resolution images are not recognized as accurately. We summarize implications for current computational models.
Collapse
Affiliation(s)
- Shichuan Du
- The Ohio State University, Columbus, OH, USA
| | | |
Collapse
|
23
|
Miellet S, Vizioli L, He L, Zhou X, Caldara R. Mapping Face Recognition Information Use across Cultures. Front Psychol 2013; 4:34. [PMID: 23430143 PMCID: PMC3576804 DOI: 10.3389/fpsyg.2013.00034] [Citation(s) in RCA: 54] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2012] [Accepted: 01/15/2013] [Indexed: 11/13/2022] Open
Abstract
Face recognition is not rooted in a universal eye movement information-gathering strategy. Western observers favor a local facial feature sampling strategy, whereas Eastern observers prefer sampling face information from a global, central fixation strategy. Yet, the precise qualitative (the diagnostic) and quantitative (the amount) information underlying these cultural perceptual biases in face recognition remains undetermined. To this end, we monitored the eye movements of Western and Eastern observers during a face recognition task, with a novel gaze-contingent technique: the Expanding Spotlight. We used 2° Gaussian apertures centered on the observers' fixations expanding dynamically at a rate of 1° every 25 ms at each fixation - the longer the fixation duration, the larger the aperture size. Identity-specific face information was only displayed within the Gaussian aperture; outside the aperture, an average face template was displayed to facilitate saccade planning. Thus, the Expanding Spotlight simultaneously maps out the facial information span at each fixation location. Data obtained with the Expanding Spotlight technique confirmed that Westerners extract more information from the eye region, whereas Easterners extract more information from the nose region. Interestingly, this quantitative difference was paired with a qualitative disparity. Retinal filters based on spatial-frequency decomposition built from the fixations maps revealed that Westerners used local high-spatial-frequency information sampling, covering all the features critical for effective face recognition (the eyes and the mouth). In contrast, Easterners achieved a similar result by using global low-spatial-frequency information from those facial features. Our data show that the face system flexibly engages into local or global eye movement strategies across cultures, by relying on distinct facial information span and culturally tuned spatially filtered information. Overall, our findings challenge the view of a unique putative process for face recognition.
Collapse
Affiliation(s)
- Sébastien Miellet
- Department of Psychology and Fribourg Center for Cognition, University of Fribourg Fribourg, Switzerland
| | | | | | | | | |
Collapse
|
24
|
Morand SM, Harvey M, Grosbras MH. Parieto-Occipital Cortex Shows Early Target Selection to Faces in a Reflexive Orienting Task. Cereb Cortex 2012. [DOI: 10.1093/cercor/bhs368] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
|
25
|
Groen IIA, Ghebreab S, Lamme VAF, Scholte HS. Spatially pooled contrast responses predict neural and perceptual similarity of naturalistic image categories. PLoS Comput Biol 2012; 8:e1002726. [PMID: 23093921 PMCID: PMC3475684 DOI: 10.1371/journal.pcbi.1002726] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2012] [Accepted: 08/02/2012] [Indexed: 11/22/2022] Open
Abstract
The visual world is complex and continuously changing. Yet, our brain transforms patterns of light falling on our retina into a coherent percept within a few hundred milliseconds. Possibly, low-level neural responses already carry substantial information to facilitate rapid characterization of the visual input. Here, we computationally estimated low-level contrast responses to computer-generated naturalistic images, and tested whether spatial pooling of these responses could predict image similarity at the neural and behavioral level. Using EEG, we show that statistics derived from pooled responses explain a large amount of variance between single-image evoked potentials (ERPs) in individual subjects. Dissimilarity analysis on multi-electrode ERPs demonstrated that large differences between images in pooled response statistics are predictive of more dissimilar patterns of evoked activity, whereas images with little difference in statistics give rise to highly similar evoked activity patterns. In a separate behavioral experiment, images with large differences in statistics were judged as different categories, whereas images with little differences were confused. These findings suggest that statistics derived from low-level contrast responses can be extracted in early visual processing and can be relevant for rapid judgment of visual similarity. We compared our results with two other, well- known contrast statistics: Fourier power spectra and higher-order properties of contrast distributions (skewness and kurtosis). Interestingly, whereas these statistics allow for accurate image categorization, they do not predict ERP response patterns or behavioral categorization confusions. These converging computational, neural and behavioral results suggest that statistics of pooled contrast responses contain information that corresponds with perceived visual similarity in a rapid, low-level categorization task. Humans excel in rapid and accurate processing of visual scenes. However, it is unclear which computations allow the visual system to convert light hitting the retina into a coherent representation of visual input in a rapid and efficient way. Here we used simple, computer-generated image categories with similar low-level structure as natural scenes to test whether a model of early integration of low-level information can predict perceived category similarity. Specifically, we show that summarized (spatially pooled) responses of model neurons covering the entire visual field (the population response) to low-level properties of visual input (contrasts) can already be informative about differences in early visual evoked activity as well as behavioral confusions of these categories. These results suggest that low-level population responses can carry relevant information to estimate similarity of controlled images, and put forward the exciting hypothesis that the visual system may exploit these responses to rapidly process real natural scenes. We propose that the spatial pooling that allows for the extraction of this information may be a plausible first step in extracting scene gist to form a rapid impression of the visual input.
Collapse
Affiliation(s)
- Iris I. A. Groen
- Cognitive Neuroscience Group, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
- * E-mail:
| | - Sennay Ghebreab
- Cognitive Neuroscience Group, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
- Intelligent Systems Lab Amsterdam, Institute of Informatics, University of Amsterdam, Amsterdam, The Netherlands
| | - Victor A. F. Lamme
- Cognitive Neuroscience Group, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
| | - H. Steven Scholte
- Cognitive Neuroscience Group, Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
| |
Collapse
|
26
|
Groen IIA, Ghebreab S, Lamme VAF, Scholte HS. Low-level contrast statistics are diagnostic of invariance of natural textures. Front Comput Neurosci 2012; 6:34. [PMID: 22701419 PMCID: PMC3370418 DOI: 10.3389/fncom.2012.00034] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2011] [Accepted: 05/23/2012] [Indexed: 11/13/2022] Open
Abstract
Texture may provide important clues for real world object and scene perception. To be reliable, these clues should ideally be invariant to common viewing variations such as changes in illumination and orientation. In a large image database of natural materials, we found textures with low-level contrast statistics that varied substantially under viewing variations, as well as textures that remained relatively constant. This led us to ask whether textures with constant contrast statistics give rise to more invariant representations compared to other textures. To test this, we selected natural texture images with either high (HV) or low (LV) variance in contrast statistics and presented these to human observers. In two distinct behavioral categorization paradigms, participants more often judged HV textures as "different" compared to LV textures, showing that textures with constant contrast statistics are perceived as being more invariant. In a separate electroencephalogram (EEG) experiment, evoked responses to single texture images (single-image ERPs) were collected. The results show that differences in contrast statistics correlated with both early and late differences in occipital ERP amplitude between individual images. Importantly, ERP differences between images of HV textures were mainly driven by illumination angle, which was not the case for LV images: there, differences were completely driven by texture membership. These converging neural and behavioral results imply that some natural textures are surprisingly invariant to illumination changes and that low-level contrast statistics are diagnostic of the extent of this invariance.
Collapse
Affiliation(s)
- Iris I A Groen
- Department of Psychology, Cognitive Neuroscience Group, University of Amsterdam Amsterdam, Netherlands
| | | | | | | |
Collapse
|
27
|
Rousselet GA, Pernet CR, Caldara R, Schyns PG. Visual Object Categorization in the Brain: What Can We Really Learn from ERP Peaks? Front Hum Neurosci 2011; 5:156. [PMID: 22144959 PMCID: PMC3228234 DOI: 10.3389/fnhum.2011.00156] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2011] [Accepted: 11/14/2011] [Indexed: 11/13/2022] Open
Affiliation(s)
- Guillaume A Rousselet
- Centre for Cognitive Neuroimaging, Institute of Neuroscience and Psychology, University of Glasgow Glasgow, UK
| | | | | | | |
Collapse
|
28
|
Gaspar CM, Rousselet GA, Pernet CR. Reliability of ERP and single-trial analyses. Neuroimage 2011; 58:620-9. [DOI: 10.1016/j.neuroimage.2011.06.052] [Citation(s) in RCA: 35] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2010] [Revised: 06/10/2011] [Accepted: 06/20/2011] [Indexed: 10/18/2022] Open
|
29
|
Rousselet GA, Gaspar CM, Wieczorek KP, Pernet CR. Modeling Single-Trial ERP Reveals Modulation of Bottom-Up Face Visual Processing by Top-Down Task Constraints (in Some Subjects). Front Psychol 2011; 2:137. [PMID: 21886627 PMCID: PMC3153882 DOI: 10.3389/fpsyg.2011.00137] [Citation(s) in RCA: 87] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2011] [Accepted: 06/09/2011] [Indexed: 11/13/2022] Open
Abstract
We studied how task constraints modulate the relationship between single-trial event-related potentials (ERPs) and image noise. Thirteen subjects performed two interleaved tasks: on different blocks, they saw the same stimuli, but they discriminated either between two faces or between two colors. Stimuli were two pictures of red or green faces that contained from 10 to 80% of phase noise, with 10% increments. Behavioral accuracy followed a noise dependent sigmoid in the identity task but was high and independent of noise level in the color task. EEG data recorded concurrently were analyzed using a single-trial ANCOVA: we assessed how changes in task constraints modulated ERP noise sensitivity while regressing out the main ERP differences due to identity, color, and task. Single-trial ERP sensitivity to image phase noise started at about 95-110 ms post-stimulus onset. Group analyses showed a significant reduction in noise sensitivity in the color task compared to the identity task from about 140 ms to 300 ms post-stimulus onset. However, statistical analyses in every subject revealed different results: significant task modulation occurred in 8/13 subjects, one showing an increase and seven showing a decrease in noise sensitivity in the color task. Onsets and durations of effects also differed between group and single-trial analyses: at any time point only a maximum of four subjects (31%) showed results consistent with group analyses. We provide detailed results for all 13 subjects, including a shift function analysis that revealed asymmetric task modulations of single-trial ERP distributions. We conclude that, during face processing, bottom-up sensitivity to phase noise can be modulated by top-down task constraints, in a broad window around the P2, at least in some subjects.
Collapse
Affiliation(s)
- Guillaume A. Rousselet
- Centre for Cognitive Neuroimaging, Institute of Neuroscience and Psychology, University of GlasgowGlasgow, UK
| | - Carl M. Gaspar
- Centre for Cognitive Neuroimaging, Institute of Neuroscience and Psychology, University of GlasgowGlasgow, UK
| | - Kacper P. Wieczorek
- Centre for Cognitive Neuroimaging, Institute of Neuroscience and Psychology, University of GlasgowGlasgow, UK
| | - Cyril R. Pernet
- Brain Research Imaging Centre, SINAPSE Collaboration, University of EdinburghEdinburgh, UK
| |
Collapse
|
30
|
Efficient bubbles for visual categorization tasks. Vision Res 2011; 51:1318-23. [DOI: 10.1016/j.visres.2011.04.007] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2010] [Revised: 04/06/2011] [Accepted: 04/07/2011] [Indexed: 11/21/2022]
|
31
|
Rousselet GA, Pernet CR. Quantifying the Time Course of Visual Object Processing Using ERPs: It's Time to Up the Game. Front Psychol 2011; 2:107. [PMID: 21779262 PMCID: PMC3132679 DOI: 10.3389/fpsyg.2011.00107] [Citation(s) in RCA: 62] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2011] [Accepted: 05/11/2011] [Indexed: 11/16/2022] Open
Abstract
Hundreds of studies have investigated the early ERPs to faces and objects using scalp and intracranial recordings. The vast majority of these studies have used uncontrolled stimuli, inappropriate designs, peak measurements, poor figures, and poor inferential and descriptive group statistics. These problems, together with a tendency to discuss any effect p < 0.05 rather than to report effect sizes, have led to a research field very much qualitative in nature, despite its quantitative inspirations, and in which predictions do not go beyond condition A > condition B. Here we describe the main limitations of face and object ERP research and suggest alternative strategies to move forward. The problems plague intracranial and surface ERP studies, but also studies using more advanced techniques – e.g., source space analyses and measurements of network dynamics, as well as many behavioral, fMRI, TMS, and LFP studies. In essence, it is time to stop amassing binary results and start using single-trial analyses to build models of visual perception.
Collapse
Affiliation(s)
- Guillaume A Rousselet
- Centre for Cognitive Neuroimaging, Institute of Neuroscience and Psychology, University of Glasgow Glasgow, UK
| | | |
Collapse
|
32
|
Schyns PG, Thut G, Gross J. Cracking the code of oscillatory activity. PLoS Biol 2011; 9:e1001064. [PMID: 21610856 PMCID: PMC3096604 DOI: 10.1371/journal.pbio.1001064] [Citation(s) in RCA: 87] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2010] [Accepted: 04/07/2011] [Indexed: 11/18/2022] Open
Abstract
Neural oscillations are ubiquitous measurements of cognitive processes and dynamic routing and gating of information. The fundamental and so far unresolved problem for neuroscience remains to understand how oscillatory activity in the brain codes information for human cognition. In a biologically relevant cognitive task, we instructed six human observers to categorize facial expressions of emotion while we measured the observers' EEG. We combined state-of-the-art stimulus control with statistical information theory analysis to quantify how the three parameters of oscillations (i.e., power, phase, and frequency) code the visual information relevant for behavior in a cognitive task. We make three points: First, we demonstrate that phase codes considerably more information (2.4 times) relating to the cognitive task than power. Second, we show that the conjunction of power and phase coding reflects detailed visual features relevant for behavioral response--that is, features of facial expressions predicted by behavior. Third, we demonstrate, in analogy to communication technology, that oscillatory frequencies in the brain multiplex the coding of visual features, increasing coding capacity. Together, our findings about the fundamental coding properties of neural oscillations will redirect the research agenda in neuroscience by establishing the differential role of frequency, phase, and amplitude in coding behaviorally relevant information in the brain.
Collapse
Affiliation(s)
- Philippe G Schyns
- Institute of Neuroscience and Psychology, University of Glasgow, Glasgow, United Kingdom.
| | | | | |
Collapse
|
33
|
Mitsudo T, Kamio Y, Goto Y, Nakashima T, Tobimatsu S. Neural responses in the occipital cortex to unrecognizable faces. Clin Neurophysiol 2010; 122:708-18. [PMID: 21071267 DOI: 10.1016/j.clinph.2010.10.004] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2010] [Revised: 09/22/2010] [Accepted: 10/08/2010] [Indexed: 11/26/2022]
Abstract
OBJECTIVE Event-related potentials (ERPs) were recorded to examine neural responses to face stimuli in a masking paradigm. METHODS Images of faces (neutral or fearful) and objects were presented in subthreshold, threshold, and suprathreshold conditions (exposure durations of approximately 20, 30 and 300 ms, respectively), followed by a 1000-ms pattern mask. We recorded ERP responses at Oz, T5, T6, Cz and Pz. The effects of physical stimulus features were examined by inverted stimuli. RESULTS The occipital N1 amplitude (approximately 160 ms) was significantly smaller in response to faces than objects when presented at a subthreshold duration. In contrast, the occipitotemporal N170 amplitude was significantly greater in the threshold and suprathreshold conditions compared with the subthreshold condition for faces, but not for objects. The P1 amplitude (approximately 120 ms) elicited by upright faces in the subthreshold condition was significantly larger than for inverted faces. CONCLUSIONS P1 and N1 components at Oz were sensitive to subthreshold faces, which suggests the presence of fast face-specific process(es) prior to face-encoding. The N170 reflects the robustness of the face selective response in the occipitotemporal area. SIGNIFICANCE Even when presented for a subthreshold duration, faces were processed differently to images of objects at an early stage of visual processing.
Collapse
Affiliation(s)
- Takako Mitsudo
- Department of Clinical Neurophysiology, Neurological Institute, Faculty of Medicine, Graduate School of Medical Sciences, Kyushu University, Higashi-ku, Fukuoka, Japan.
| | | | | | | | | |
Collapse
|
34
|
Daniel S, Bentin S. Age-related changes in processing faces from detection to identification: ERP evidence. Neurobiol Aging 2010; 33:206.e1-28. [PMID: 20961658 DOI: 10.1016/j.neurobiolaging.2010.09.001] [Citation(s) in RCA: 49] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2010] [Revised: 08/19/2010] [Accepted: 09/05/2010] [Indexed: 11/27/2022]
Abstract
We examined the ability of people 70 to 90 years old to apply global, configural, and featural face-processing strategies. In addition we investigated age-related changes in the ability to categorize faces at basic, subordinate, and individual levels. Using the N170 potential as index of early face processing and the P300 component as index of categorical decision making and effort, we found significant age-related perceptual changes which slowed and somewhat impaired face processing. Specifically, older participants had problems integrating face features into global structures, demonstrating enhanced dependence on distal global information. They did not apply configural computations by default while processing faces which suggests that, unless identification is required, they process faces only at a basic level. These perceptual changes could be the cause for slower and less accurate subordinate categorization, particularly when it is based on details. At the neural levels face processing was not right-lateralized, reflecting excessive involvement of the left hemisphere in perception leading to a more general reduction of interhemispheric asymmetry. In addition we found excessive but nonselective activation of frontal regions adding support to the view that executive control and particularly inhibition of irrelevant input are reduced in the elderly.
Collapse
Affiliation(s)
- Sharon Daniel
- Department of Neurobiology, The Hebrew University of Jerusalem, Jerusalem, Israel
| | | |
Collapse
|
35
|
Transmission of facial expressions of emotion co-evolved with their efficient decoding in the brain: behavioral and brain evidence. PLoS One 2009; 4:e5625. [PMID: 19462006 PMCID: PMC2680487 DOI: 10.1371/journal.pone.0005625] [Citation(s) in RCA: 72] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2008] [Accepted: 04/10/2009] [Indexed: 11/19/2022] Open
Abstract
Competent social organisms will read the social signals of their peers. In primates, the face has evolved to transmit the organism's internal emotional state. Adaptive action suggests that the brain of the receiver has co-evolved to efficiently decode expression signals. Here, we review and integrate the evidence for this hypothesis. With a computational approach, we co-examined facial expressions as signals for data transmission and the brain as receiver and decoder of these signals. First, we show in a model observer that facial expressions form a lowly correlated signal set. Second, using time-resolved EEG data, we show how the brain uses spatial frequency information impinging on the retina to decorrelate expression categories. Between 140 to 200 ms following stimulus onset, independently in the left and right hemispheres, an information processing mechanism starts locally with encoding the eye, irrespective of expression, followed by a zooming out to processing the entire face, followed by a zooming back in to diagnostic features (e.g. the opened eyes in “fear”, the mouth in “happy”). A model categorizer demonstrates that at 200 ms, the left and right brain have represented enough information to predict behavioral categorization performance.
Collapse
|