1
|
Neurocognitive effects of a training program for poor face recognizers using shape and texture caricatures: A pilot investigation. Neuropsychologia 2021; 165:108133. [PMID: 34971671 DOI: 10.1016/j.neuropsychologia.2021.108133] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2020] [Revised: 12/09/2021] [Accepted: 12/19/2021] [Indexed: 12/30/2022]
Abstract
Recent research suggested disproportional usage of shape information by people with poor face recognition, although texture information appears to be more important for familiar face recognition. Here, we tested a training program with faces that were selectively caricatured in either shape or texture parameters. Forty-eight young adults with poor face recognition skills (1 SD below the mean in at least 2/3 face processing tests: CFMT, GFMT, BFFT) were pseudo-randomly assigned to either one of two training groups or a control group (n = 16 each). Training comprised six sessions over three weeks. Per session, participants studied ten unfamiliar facial identities whose shape or texture characteristics were caricatured. Before and after training (or waiting in the control group), all participants completed EEG experiments on face learning and famous face recognition, and behavioral face processing tests. Results showed small but specific training-induced improvements: Whereas shape training improved face matching (training tasks, and to some extent GFMT), texture training elicited marked improvements in face learning (CFMT). Moreover, for the texture training group the N170 ERP was enhanced for novel faces post-training, suggesting training-induced changes in early markers of face processing. Although further research is necessary, this suggests that parameter-specific caricature training is a promising way to improve performance in people with poor face recognition skills.
Collapse
|
2
|
Zheng Y, Jia S, Yu Z, Liu JK, Huang T. Unraveling neural coding of dynamic natural visual scenes via convolutional recurrent neural networks. PATTERNS (NEW YORK, N.Y.) 2021; 2:100350. [PMID: 34693375 PMCID: PMC8515013 DOI: 10.1016/j.patter.2021.100350] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/03/2021] [Revised: 06/22/2021] [Accepted: 08/23/2021] [Indexed: 11/18/2022]
Abstract
Traditional models of retinal system identification analyze the neural response to artificial stimuli using models consisting of predefined components. The model design is limited to prior knowledge, and the artificial stimuli are too simple to be compared with stimuli processed by the retina. To fill in this gap with an explainable model that reveals how a population of neurons work together to encode the larger field of natural scenes, here we used a deep-learning model for identifying the computational elements of the retinal circuit that contribute to learning the dynamics of natural scenes. Experimental results verify that the recurrent connection plays a key role in encoding complex dynamic visual scenes while learning biological computational underpinnings of the retinal circuit. In addition, the proposed models reveal both the shapes and the locations of the spatiotemporal receptive fields of ganglion cells.
Collapse
Affiliation(s)
- Yajing Zheng
- Department of Computer Science and Technology, National Engineering Laboratory for Video Technology, Peking University, Beijing 100871, China
| | - Shanshan Jia
- Department of Computer Science and Technology, National Engineering Laboratory for Video Technology, Peking University, Beijing 100871, China
| | - Zhaofei Yu
- Department of Computer Science and Technology, National Engineering Laboratory for Video Technology, Peking University, Beijing 100871, China
- Institute for Artificial Intelligence, Peking University, Beijing 100871, China
| | - Jian K. Liu
- School of Computing, University of Leeds, Leeds LS2 9JT, UK
| | - Tiejun Huang
- Department of Computer Science and Technology, National Engineering Laboratory for Video Technology, Peking University, Beijing 100871, China
- Institute for Artificial Intelligence, Peking University, Beijing 100871, China
| |
Collapse
|
3
|
Spoerer CJ, Kietzmann TC, Mehrer J, Charest I, Kriegeskorte N. Recurrent neural networks can explain flexible trading of speed and accuracy in biological vision. PLoS Comput Biol 2020; 16:e1008215. [PMID: 33006992 PMCID: PMC7556458 DOI: 10.1371/journal.pcbi.1008215] [Citation(s) in RCA: 45] [Impact Index Per Article: 11.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2019] [Revised: 10/14/2020] [Accepted: 08/03/2020] [Indexed: 11/18/2022] Open
Abstract
Deep feedforward neural network models of vision dominate in both computational neuroscience and engineering. The primate visual system, by contrast, contains abundant recurrent connections. Recurrent signal flow enables recycling of limited computational resources over time, and so might boost the performance of a physically finite brain or model. Here we show: (1) Recurrent convolutional neural network models outperform feedforward convolutional models matched in their number of parameters in large-scale visual recognition tasks on natural images. (2) Setting a confidence threshold, at which recurrent computations terminate and a decision is made, enables flexible trading of speed for accuracy. At a given confidence threshold, the model expends more time and energy on images that are harder to recognise, without requiring additional parameters for deeper computations. (3) The recurrent model's reaction time for an image predicts the human reaction time for the same image better than several parameter-matched and state-of-the-art feedforward models. (4) Across confidence thresholds, the recurrent model emulates the behaviour of feedforward control models in that it achieves the same accuracy at approximately the same computational cost (mean number of floating-point operations). However, the recurrent model can be run longer (higher confidence threshold) and then outperforms parameter-matched feedforward comparison models. These results suggest that recurrent connectivity, a hallmark of biological visual systems, may be essential for understanding the accuracy, flexibility, and dynamics of human visual recognition.
Collapse
Affiliation(s)
- Courtney J. Spoerer
- Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, United Kingdom
| | - Tim C. Kietzmann
- Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, United Kingdom
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
| | - Johannes Mehrer
- Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge, United Kingdom
| | - Ian Charest
- School of Psychology and Centre for Human Brain Health, University of Birmingham, United Kingdom
| | - Nikolaus Kriegeskorte
- Department of Psychology, Department of Neuroscience, Department of Electrical Engineering, Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USA
| |
Collapse
|
4
|
Orlandi A, Proverbio AM. Bilateral engagement of the occipito-temporal cortex in response to dance kinematics in experts. Sci Rep 2019; 9:1000. [PMID: 30700799 PMCID: PMC6353946 DOI: 10.1038/s41598-018-37876-x] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2018] [Accepted: 12/14/2018] [Indexed: 01/04/2023] Open
Abstract
Previous evidence has shown neuroplastic changes in brain anatomy and connectivity associated with the acquisition of professional visuomotor skills. Reduced hemispherical asymmetry was found in the sensorimotor and visual areas in expert musicians and athletes compared with non-experts. Moreover, increased expertise with faces, body, and objects resulted in an enhanced engagement of the occipito-temporal cortex (OTC) during stimulus observation. The present study aimed at investigating whether intense and extended practice with dance would result in an enhanced symmetric response of OTC at an early stage of action processing. Expert ballet dancers and non-dancer controls were presented with videos depicting ballet steps during EEG recording. The observation of the moving dancer elicited a posterior N2 component, being larger over the left hemisphere in dancers than controls. The source reconstruction (swLORETA) of the negativity showed the engagement of the bilateral inferior and middle temporal regions in experts, while right-lateralized activity was found in controls. The dancers also showed an early P2 and enhanced P300 responses, indicating faster stimulus processing and subsequent recognition. This evidence seemed to suggest expertise-related increased sensitivity of the OTC in encoding body kinematics. Thus, we speculated that long-term whole-body practice would result in enriched and refined action processing.
Collapse
Affiliation(s)
- Andrea Orlandi
- Neuro-MI, Milan Center for Neuroscience, Department of Psychology, University of Milano - Bicocca, Milan, Italy.
| | - Alice Mado Proverbio
- Neuro-MI, Milan Center for Neuroscience, Department of Psychology, University of Milano - Bicocca, Milan, Italy
| |
Collapse
|
5
|
Akhavein H, Dehmoobadsharifabadi A, Farivar R. Magnetoencephalography adaptation reveals depth-cue-invariant object representations in the visual cortex. J Vis 2018; 18:6. [PMID: 30458514 DOI: 10.1167/18.12.6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Independent of edges and 2-D shape that can be highly informative of object identity, depth cues alone can also give rise to vivid and effective object percepts. The processing of different depth cues engages segregated cortical areas, and an efficient object representation would be one that is invariant to depth cues. Here, we investigated depth-cue invariance of object representations by measuring the category-specific response to faces-the M170 response measured with magnetoencephalography. The M170 response is strongest to faces and is sensitive to adaptation, such that repeated presentation of a face diminishes subsequent M170 responses. We used this feature of the M170 and measured the degree to which the adaptation effect is affected by variations in depth cue and 3-D object shape. Subjects viewed a rapid presentation of two stimuli-an adaptor and a test stimulus. The adaptor was either a face, a chair, or a face-like oval surface, and rendered with a single depth cue (shading, structure from motion, or texture). The test stimulus was always a shaded face of a random identity, thus completely controlling for low-level influences on the M170 response to the test stimulus. In the left fusiform face area, we found strong M170 adaptation when the adaptor was a face regardless of its depth cue. This adaptation was marginal in the right fusiform and negligible in the occipital regions. Our results support the presence of depth-cue-invariant representations in the human visual system, alongside size, position, and viewpoint invariance.
Collapse
Affiliation(s)
- Hassan Akhavein
- McGill Vision Research, Department of Ophthalmology, McGill University, Montreal, Canada
| | | | - Reza Farivar
- McGill Vision Research, Department of Ophthalmology, McGill University, Montreal, Canada
| |
Collapse
|
6
|
Limbach K, Kaufmann JM, Wiese H, Witte OW, Schweinberger SR. Enhancement of face-sensitive ERPs in older adults induced by face recognition training. Neuropsychologia 2018; 119:197-213. [PMID: 30114386 DOI: 10.1016/j.neuropsychologia.2018.08.010] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/07/2018] [Revised: 08/01/2018] [Accepted: 08/11/2018] [Indexed: 12/30/2022]
Abstract
A common cognitive problem reported by older people is compromised face recognition, which is often paralleled by age-related changes in face-sensitive and memory-related components in event-related brain potentials (ERPs). We developed a new training using photorealistic caricatures based on evidence that caricatures are beneficial for people with compromised face processing. Twenty-four older participants (62-75 yrs, 13 female) completed 12 training sessions (3 per week, 60 min each) and 24 older participants (61-76 yrs, 12 female) acted as controls. Before and after training (or waiting), participants took part in a diagnostic test battery for face processing abilities, and in ERP experiments on face learning and recognition. Although performance improvements during the training provided little evidence for generalization to other face processing tasks, ERPs showed substantial training-related enhancements of face-sensitive ERPs. Specifically, we observed marked increases of the N170, P200 and N250 components, which may indicate training-induced enhancement of face detection and activation of identity-specific representations. Thus, neuronal correlates of face processing are plastic in older age, and can be modulated by caricature training.
Collapse
Affiliation(s)
- Katharina Limbach
- Department of General Psychology, Friedrich Schiller University, Jena, Germany.
| | - Jürgen M Kaufmann
- Department of General Psychology, Friedrich Schiller University, Jena, Germany; DFG Research Unit Person Perception, Jena, Germany
| | - Holger Wiese
- Department of Psychology, Durham University, UK; DFG Research Unit Person Perception, Jena, Germany
| | | | - Stefan R Schweinberger
- Department of General Psychology, Friedrich Schiller University, Jena, Germany; DFG Research Unit Person Perception, Jena, Germany.
| |
Collapse
|