1
|
Talwar A, Cormack F, Huys QJM, Roiser JP. A hierarchical reinforcement learning model explains individual differences in attentional set shifting. COGNITIVE, AFFECTIVE & BEHAVIORAL NEUROSCIENCE 2024; 24:1008-1022. [PMID: 39313748 PMCID: PMC11525250 DOI: 10.3758/s13415-024-01223-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 08/27/2024] [Indexed: 09/25/2024]
Abstract
Attentional set shifting refers to the ease with which the focus of attention is directed and switched. Cognitive tasks, such as the widely used CANTAB IED, reveal great variation in set shifting ability in the general population, with notable impairments in those with psychiatric diagnoses. The attentional and learning processes underlying this cognitive ability and how they lead to the observed variation remain unknown. To directly test this, we used a modelling approach on two independent large-scale online general-population samples performing CANTAB IED, with one including additional psychiatric symptom assessment. We found a hierarchical model that learnt both feature values and dimension attention best explained the data and that compulsive symptoms were associated with slower learning and higher attentional bias to the first relevant stimulus dimension. These data showcase a new methodology to analyse data from the CANTAB IED task, as well as suggest a possible mechanistic explanation for the variation in set shifting performance, and its relationship to compulsive symptoms.
Collapse
Affiliation(s)
- Anahita Talwar
- Neuroscience and Mental Health Group, UCL Institute of Cognitive Neuroscience, 17-19 Queen Square, London, WC1N 3AZ, UK
- Cambridge Cognition Ltd, Tunbridge Court, Bottisham, Cambridge, CB25 9TU, UK
| | - Francesca Cormack
- Cambridge Cognition Ltd, Tunbridge Court, Bottisham, Cambridge, CB25 9TU, UK
| | - Quentin J M Huys
- Applied Computational Psychiatry Lab, Mental Health Neuroscience Department, Division of Psychiatry and Max Planck Centre for Computational Psychiatry and Ageing Research, Queen Square Institute of Neurology, UCL, Maple House, 149 Tottenham Court Rd, London, W1T 7BN, UK
| | - Jonathan P Roiser
- Neuroscience and Mental Health Group, UCL Institute of Cognitive Neuroscience, 17-19 Queen Square, London, WC1N 3AZ, UK.
| |
Collapse
|
2
|
Song Y, Wang Q, Fang F. Time courses of brain plasticity underpinning visual motion perceptual learning. Neuroimage 2024; 302:120897. [PMID: 39442899 DOI: 10.1016/j.neuroimage.2024.120897] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 10/10/2024] [Accepted: 10/21/2024] [Indexed: 10/25/2024] Open
Abstract
Visual perceptual learning (VPL) refers to a long-term improvement of visual task performance through training or experience, reflecting brain plasticity even in adults. In human subjects, VPL has been mostly studied using functional magnetic resonance imaging (fMRI). However, due to the low temporal resolution of fMRI, how VPL affects the time course of visual information processing is largely unknown. To address this issue, we trained human subjects to perform a visual motion direction discrimination task. Their behavioral performance and magnetoencephalography (MEG) signals responding to the motion stimuli were measured before, immediately after, and two weeks after training. Training induced a long-lasting behavioral improvement for the trained direction. Based on the MEG signals from occipital sensors, we found that, for the trained motion direction, VPL increased the motion direction decoding accuracy, reduced the motion direction decoding latency, enhanced the direction-selective channel response, and narrowed the tuning profile. Following the MEG source reconstruction, we showed that VPL enhanced the cortical response in early visual cortex (EVC) and strengthened the feedforward connection from EVC to V3A. These VPL-induced neural changes co-occurred in 160-230 ms after stimulus onset. Complementary to previous fMRI findings on VPL, this study provides a comprehensive description on the neural mechanisms of visual motion perceptual learning from a temporal perspective and reveals how VPL shapes the time course of visual motion processing in the adult human brain.
Collapse
Affiliation(s)
- Yongqian Song
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing 100871, China; IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China; Peking-Tsinghua Center for Life Sciences, Peking University, Beijing 100871, China
| | - Qian Wang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing 100871, China; IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China; National Key Laboratory of General Artificial Intelligence, Peking University, Beijing 100871, China
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing 100871, China; IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China; Peking-Tsinghua Center for Life Sciences, Peking University, Beijing 100871, China; Key Laboratory of Machine Perception (Ministry of Education), Peking University, Beijing 100871, China.
| |
Collapse
|
3
|
Wenger M, Maimon A, Yizhar O, Snir A, Sasson Y, Amedi A. Hearing temperatures: employing machine learning for elucidating the cross-modal perception of thermal properties through audition. Front Psychol 2024; 15:1353490. [PMID: 39156805 PMCID: PMC11327021 DOI: 10.3389/fpsyg.2024.1353490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2023] [Accepted: 06/28/2024] [Indexed: 08/20/2024] Open
Abstract
People can use their sense of hearing for discerning thermal properties, though they are for the most part unaware that they can do so. While people unequivocally claim that they cannot perceive the temperature of pouring water through the auditory properties of hearing it being poured, our research further strengthens the understanding that they can. This multimodal ability is implicitly acquired in humans, likely through perceptual learning over the lifetime of exposure to the differences in the physical attributes of pouring water. In this study, we explore people's perception of this intriguing cross modal correspondence, and investigate the psychophysical foundations of this complex ecological mapping by employing machine learning. Our results show that not only can the auditory properties of pouring water be classified by humans in practice, the physical characteristics underlying this phenomenon can also be classified by a pre-trained deep neural network.
Collapse
Affiliation(s)
- Mohr Wenger
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
- Department of Cognitive and Brain Sciences, The Hebrew University of Jerusalem, Jerusalem, Israel
| | - Amber Maimon
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
- Computational Psychiatry and Neurotechnology Lab, Department of Brain and Cognitive Sciences, Ben Gurion University, Be’er Sheva, Israel
| | - Or Yizhar
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
- Department of Cognitive and Brain Sciences, The Hebrew University of Jerusalem, Jerusalem, Israel
- Research Group Adaptive Memory and Decision Making, Max Planck Institute for Human Development, Berlin, Germany
| | - Adi Snir
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
| | - Yonatan Sasson
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
| | - Amir Amedi
- Baruch Ivcher Institute for Brain Cognition and Technology, Baruch Ivcher School of Psychology, Reichman University, Herzliya, Israel
| |
Collapse
|
4
|
Zhu JP, Zhang JY. Feature variability determines specificity and transfer in multiorientation feature detection learning. J Vis 2024; 24:2. [PMID: 38691087 PMCID: PMC11079675 DOI: 10.1167/jov.24.5.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Accepted: 02/26/2024] [Indexed: 05/03/2024] Open
Abstract
Historically, in many perceptual learning experiments, only a single stimulus is practiced, and learning is often specific to the trained feature. Our prior work has demonstrated that multi-stimulus learning (e.g., training-plus-exposure procedure) has the potential to achieve generalization. Here, we investigated two important characteristics of multi-stimulus learning, namely, roving and feature variability, and their impacts on multi-stimulus learning and generalization. We adopted a feature detection task in which an oddly oriented target bar differed by 16° from the background bars. The stimulus onset asynchrony threshold between the target and the mask was measured with a staircase procedure. Observers were trained with four target orientation search stimuli, either with a 5° deviation (30°-35°-40°-45°) or with a 45° deviation (30°-75°-120°-165°), and the four reference stimuli were presented in a roving manner. The transfer of learning to the swapped target-background orientations was evaluated after training. We found that multi-stimulus training with a 5° deviation resulted in significant learning improvement, but learning failed to transfer to the swapped target-background orientations. In contrast, training with a 45° deviation slowed learning but produced a significant generalization to swapped orientations. Furthermore, a modified training-plus-exposure procedure, in which observers were trained with four orientation search stimuli with a 5° deviation and simultaneously passively exposed to orientations with high feature variability (45° deviation), led to significant orientation learning generalization. Learning transfer also occurred when the four orientation search stimuli with a 5° deviation were presented in separate blocks. These results help us to specify the condition under which multistimuli learning produces generalization, which holds potential for real-world applications of perceptual learning, such as vision rehabilitation and expert training.
Collapse
Affiliation(s)
- Jun-Ping Zhu
- School of Psychological and Cognitive Sciences, and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
| | - Jun-Yun Zhang
- School of Psychological and Cognitive Sciences, and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, China
| |
Collapse
|
5
|
Shen S, Sun Y, Lu J, Li C, Chen Q, Mo C, Fang F, Zhang X. Profiles of visual perceptual learning in feature space. iScience 2024; 27:109128. [PMID: 38384835 PMCID: PMC10879700 DOI: 10.1016/j.isci.2024.109128] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 01/22/2024] [Accepted: 02/01/2024] [Indexed: 02/23/2024] Open
Abstract
Visual perceptual learning (VPL), experience-induced gains in discriminating visual features, has been studied extensively and intensively for many years, its profile in feature space, however, remains unclear. Here, human subjects were trained to perform either a simple low-level feature (grating orientation) or a complex high-level object (face view) discrimination task over a long-time course. During, immediately after, and one month after training, all results showed that in feature space VPL in grating orientation discrimination was a center-surround profile; VPL in face view discrimination, however, was a monotonic gradient profile. Importantly, these two profiles can be emerged by a deep convolutional neural network with a modified AlexNet consisted of 7 and 12 layers, respectively. Altogether, our study reveals for the first time a feature hierarchy-dependent profile of VPL in feature space, placing a necessary constraint on our understanding of the neural computation of VPL.
Collapse
Affiliation(s)
- Shiqi Shen
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| | - Yueling Sun
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| | - Jiachen Lu
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| | - Chu Li
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| | - Qinglin Chen
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| | - Ce Mo
- Department of Psychology, Sun-YatSen University, Guangzhou, Guangdong 510275, China
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing 100871, China
- IDG/McGovern Institute for Brain Research, Peking University, Beijing 100871, China
- Peking-Tsinghua Center for Life Sciences, Peking University, Beijing 100871, China
| | - Xilin Zhang
- Key Laboratory of Brain, Cognition and Education Sciences, Ministry of Education, South China Normal University, Guangzhou, Guangdong 510631, China
- School of Psychology, Center for Studies of Psychological Application, and Guangdong Provincial Key Laboratory of Mental Health and Cognitive Science, South China Normal University, Guangzhou, Guangdong 510631, China
| |
Collapse
|
6
|
Laamerad P, Awada A, Pack CC, Bakhtiari S. Asymmetric stimulus representations bias visual perceptual learning. J Vis 2024; 24:10. [PMID: 38285454 PMCID: PMC10829801 DOI: 10.1167/jov.24.1.10] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2023] [Accepted: 12/12/2023] [Indexed: 01/30/2024] Open
Abstract
The primate visual cortex contains various regions that exhibit specialization for different stimulus properties, such as motion, shape, and color. Within each region, there is often further specialization, such that particular stimulus features, such as horizontal and vertical orientations, are over-represented. These asymmetries are associated with well-known perceptual biases, but little is known about how they influence visual learning. Most theories would predict that learning is optimal, in the sense that it is unaffected by these asymmetries. However, other approaches to learning would result in specific patterns of perceptual biases. To distinguish between these possibilities, we trained human observers to discriminate between expanding and contracting motion patterns, which have a highly asymmetrical representation in the visual cortex. Observers exhibited biased percepts of these stimuli, and these biases were affected by training in ways that were often suboptimal. We simulated different neural network models and found that a learning rule that involved only adjustments to decision criteria, rather than connection weights, could account for our data. These results suggest that cortical asymmetries influence visual perception and that human observers often rely on suboptimal strategies for learning.
Collapse
Affiliation(s)
- Pooya Laamerad
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill University, Montreal, Canada
| | - Asmara Awada
- Department of Psychology, Université de Montréal, Montreal, Canada
| | - Christopher C Pack
- Department of Neurology and Neurosurgery, Montreal Neurological Institute, McGill University, Montreal, Canada
| | - Shahab Bakhtiari
- Department of Psychology, Université de Montréal, Montreal, Canada
- Mila - Quebec AI Institute, Montreal, Canada
| |
Collapse
|
7
|
Veerabadran V, Goldman J, Shankar S, Cheung B, Papernot N, Kurakin A, Goodfellow I, Shlens J, Sohl-Dickstein J, Mozer MC, Elsayed GF. Subtle adversarial image manipulations influence both human and machine perception. Nat Commun 2023; 14:4933. [PMID: 37582834 PMCID: PMC10427626 DOI: 10.1038/s41467-023-40499-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2021] [Accepted: 08/01/2023] [Indexed: 08/17/2023] Open
Abstract
Although artificial neural networks (ANNs) were inspired by the brain, ANNs exhibit a brittleness not generally observed in human perception. One shortcoming of ANNs is their susceptibility to adversarial perturbations-subtle modulations of natural images that result in changes to classification decisions, such as confidently mislabelling an image of an elephant, initially classified correctly, as a clock. In contrast, a human observer might well dismiss the perturbations as an innocuous imaging artifact. This phenomenon may point to a fundamental difference between human and machine perception, but it drives one to ask whether human sensitivity to adversarial perturbations might be revealed with appropriate behavioral measures. Here, we find that adversarial perturbations that fool ANNs similarly bias human choice. We further show that the effect is more likely driven by higher-order statistics of natural images to which both humans and ANNs are sensitive, rather than by the detailed architecture of the ANN.
Collapse
Affiliation(s)
- Vijay Veerabadran
- Google, Mountain View, CA, USA
- Department of Cognitive Science, University of California, San Diego, CA, USA
| | | | - Shreya Shankar
- Google, Mountain View, CA, USA
- University of California, Berkeley, CA, USA
| | - Brian Cheung
- Google, Mountain View, CA, USA
- MIT Brain and Cognitive Sciences, Cambridge, MA, USA
| | | | | | | | | | | | | | | |
Collapse
|
8
|
Manenti GL, Dizaji AS, Schwiedrzik CM. Variability in training unlocks generalization in visual perceptual learning through invariant representations. Curr Biol 2023; 33:817-826.e3. [PMID: 36724782 DOI: 10.1016/j.cub.2023.01.011] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2022] [Revised: 12/24/2022] [Accepted: 01/06/2023] [Indexed: 02/03/2023]
Abstract
Stimulus and location specificity are long considered hallmarks of visual perceptual learning. This renders visual perceptual learning distinct from other forms of learning, where generalization can be more easily attained, and therefore unsuitable for practical applications, where generalization is key. Based on the hypotheses derived from the structure of the visual system, we test here whether stimulus variability can unlock generalization in perceptual learning. We train subjects in orientation discrimination, while we vary the amount of variability in a task-irrelevant feature, spatial frequency. We find that, independently of task difficulty, this manipulation enables generalization of learning to new stimuli and locations, while not negatively affecting the overall amount of learning on the task. We then use deep neural networks to investigate how variability unlocks generalization. We find that networks develop invariance to the task-irrelevant feature when trained with variable inputs. The degree of learned invariance strongly predicts generalization. A reliance on invariant representations can explain variability-induced generalization in visual perceptual learning. This suggests new targets for understanding the neural basis of perceptual learning in the higher-order visual cortex and presents an easy-to-implement modification of common training paradigms that may benefit practical applications.
Collapse
Affiliation(s)
- Giorgio L Manenti
- Neural Circuits and Cognition Lab, European Neuroscience Institute Göttingen, A Joint Initiative of the University Medical Center Göttingen and the Max Planck Society, Grisebachstraße 5, 37077 Göttingen, Germany; Perception and Plasticity Group, German Primate Center, Leibniz Institute for Primate Research, Kellnerweg 4, 37077 Göttingen, Germany; Systems Neuroscience Program, Graduate School for Neurosciences, Biophysics and Molecular Biosciences (GGNB), 37077 Göttingen, Germany
| | - Aslan S Dizaji
- Neural Circuits and Cognition Lab, European Neuroscience Institute Göttingen, A Joint Initiative of the University Medical Center Göttingen and the Max Planck Society, Grisebachstraße 5, 37077 Göttingen, Germany
| | - Caspar M Schwiedrzik
- Neural Circuits and Cognition Lab, European Neuroscience Institute Göttingen, A Joint Initiative of the University Medical Center Göttingen and the Max Planck Society, Grisebachstraße 5, 37077 Göttingen, Germany; Perception and Plasticity Group, German Primate Center, Leibniz Institute for Primate Research, Kellnerweg 4, 37077 Göttingen, Germany.
| |
Collapse
|
9
|
Vogels R. Perceptual learning: Breaking specificity by variability. Curr Biol 2023; 33:R182-R185. [PMID: 36917939 DOI: 10.1016/j.cub.2023.01.025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/16/2023]
Abstract
A new psychophysical study in humans suggests that increasing the variability of task-irrelevant features during training enhances the generalization of visual perceptual learning to untrained stimuli and locations.
Collapse
Affiliation(s)
- Rufin Vogels
- Leuven Brain Institute, KU Leuven, Leuven, Belgium.
| |
Collapse
|
10
|
Benjamin AS, Zhang LQ, Qiu C, Stocker AA, Kording KP. Efficient neural codes naturally emerge through gradient descent learning. Nat Commun 2022; 13:7972. [PMID: 36581618 PMCID: PMC9800366 DOI: 10.1038/s41467-022-35659-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 12/14/2022] [Indexed: 12/30/2022] Open
Abstract
Human sensory systems are more sensitive to common features in the environment than uncommon features. For example, small deviations from the more frequently encountered horizontal orientations can be more easily detected than small deviations from the less frequent diagonal ones. Here we find that artificial neural networks trained to recognize objects also have patterns of sensitivity that match the statistics of features in images. To interpret these findings, we show mathematically that learning with gradient descent in neural networks preferentially creates representations that are more sensitive to common features, a hallmark of efficient coding. This effect occurs in systems with otherwise unconstrained coding resources, and additionally when learning towards both supervised and unsupervised objectives. This result demonstrates that efficient codes can naturally emerge from gradient-like learning.
Collapse
Affiliation(s)
- Ari S Benjamin
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, USA.
| | - Ling-Qi Zhang
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, USA
| | - Cheng Qiu
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, USA
| | - Alan A Stocker
- Department of Psychology, University of Pennsylvania, Philadelphia, PA, USA
| | - Konrad P Kording
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, USA
- Department of Neuroscience, University of Pennsylvania, Philadelphia, PA, USA
| |
Collapse
|
11
|
Shan H, Sompolinsky H. Minimum perturbation theory of deep perceptual learning. Phys Rev E 2022; 106:064406. [PMID: 36671118 DOI: 10.1103/physreve.106.064406] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2022] [Accepted: 11/22/2022] [Indexed: 06/17/2023]
Abstract
Perceptual learning (PL) involves long-lasting improvement in perceptual tasks following extensive training and is accompanied by modified neuronal responses in sensory cortical areas in the brain. Understanding the dynamics of PL and the resultant synaptic changes is important for causally connecting PL to the observed neural plasticity. This is theoretically challenging because learning-related changes are distributed across many stages of the sensory hierarchy. In this paper, we modeled the sensory hierarchy as a deep nonlinear neural network and studied PL of fine discrimination, a common and well-studied paradigm of PL. Using tools from statistical physics, we developed a mean-field theory of the network in the limit of a large number of neurons and large number of examples. Our theory suggests that, in this thermodynamic limit, the input-output function of the network can be exactly mapped to that of a deep linear network, allowing us to characterize the space of solutions for the task. Surprisingly, we found that modifying synaptic weights in the first layer of the hierarchy is both sufficient and necessary for PL. To address the degeneracy of the space of solutions, we postulate that PL dynamics are constrained by a normative minimum perturbation (MP) principle, which favors weight matrices with minimal changes relative to their prelearning values. Interestingly, MP plasticity induces changes to weights and neural representations in all layers of the network, except for the readout weight vector. While weight changes in higher layers are not necessary for learning, they help reduce overall perturbation to the network. In addition, such plasticity can be learned simply through slow learning. We further elucidate the properties of MP changes and compare them against experimental findings. Overall, our statistical mechanics theory of PL provides mechanistic and normative understanding of several important empirical findings of PL.
Collapse
Affiliation(s)
- Haozhe Shan
- Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138, USA and Program in Neuroscience, Harvard Medical School, Boston, Massachusetts 02115, USA
| | - Haim Sompolinsky
- Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138, USA and Edmond and Lily Safra Center for Brain Sciences, Hebrew University of Jerusalem, Jerusalem 9190401, Israel
| |
Collapse
|
12
|
Turnbull A, Seitz A, Tadin D, Lin FV. Unifying framework for cognitive training interventions in brain aging. Ageing Res Rev 2022; 81:101724. [PMID: 36031055 PMCID: PMC10681332 DOI: 10.1016/j.arr.2022.101724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2022] [Revised: 07/29/2022] [Accepted: 08/22/2022] [Indexed: 01/31/2023]
Abstract
Cognitive training is a promising tool for slowing or preventing cognitive decline in older adults at-risk for dementia. Its success, however, has been limited by a lack of evidence showing that it reliably causes broad training effects: improvements in cognition across a range of domains that lead to real-world benefits. Here, we propose a framework for enhancing the effect of cognitive training interventions in brain aging. The focus is on (A) developing cognitive training task paradigms that are informed by population-level cognitive characteristics and pathophysiology, and (B) personalizing how these sets are presented to participants during training via feedback loops that aim to optimize "mismatch" between participant capacity and training demands using both adaptation and random variability. In this way, cognitive training can better alter whole-brain topology in a manner that supports broad training effects in the context of brain aging.
Collapse
Affiliation(s)
- Adam Turnbull
- University of Rochester, USA; Stanford University, USA
| | | | | | | |
Collapse
|
13
|
Lu ZL, Dosher BA. Current directions in visual perceptual learning. NATURE REVIEWS PSYCHOLOGY 2022; 1:654-668. [PMID: 37274562 PMCID: PMC10237053 DOI: 10.1038/s44159-022-00107-2] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 08/16/2022] [Indexed: 06/06/2023]
Abstract
The visual expertise of adult humans is jointly determined by evolution, visual development, and visual perceptual learning. Perceptual learning refers to performance improvements in perceptual tasks after practice or training in the task. It occurs in almost all visual tasks, ranging from simple feature detection to complex scene analysis. In this Review, we focus on key behavioral aspects of visual perceptual learning. We begin by describing visual perceptual learning tasks and manipulations that influence the magnitude of learning, and then discuss specificity of learning. Next, we present theories and computational models of learning and specificity. We then review applications of visual perceptual learning in visual rehabilitation. Finally, we summarize the general principles of visual perceptual learning, discuss the tension between plasticity and stability, and conclude with new research directions.
Collapse
Affiliation(s)
- Zhong-Lin Lu
- Division of Arts and Sciences, New York University Shanghai, Shanghai, China
- Center for Neural Science, New York University, New York, NY, USA
- Department of Psychology, New York University, New York, NY, USA
- Institute of Brain and Cognitive Science, New York University - East China Normal University, Shanghai, China
| | | |
Collapse
|
14
|
Tao P, Cheng J, Chen L. Brain-inspired chaotic backpropagation for MLP. Neural Netw 2022; 155:1-13. [PMID: 36027661 DOI: 10.1016/j.neunet.2022.08.004] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Revised: 06/14/2022] [Accepted: 08/03/2022] [Indexed: 11/17/2022]
Abstract
Backpropagation (BP) algorithm is one of the most basic learning algorithms in deep learning. Although BP has been widely used, it still suffers from the problem of easily falling into the local minima due to its gradient dynamics. Inspired by the fact that the learning of real brains may exploit chaotic dynamics, we propose the chaotic backpropagation (CBP) algorithm by integrating the intrinsic chaos of real neurons into BP. By validating on multiple datasets (e.g. cifar10), we show that, for multilayer perception (MLP), CBP has significantly better abilities than those of BP and its variants in terms of optimization and generalization from both computational and theoretical viewpoints. Actually, CBP can be regarded as a general form of BP with global searching ability inspired by the chaotic learning process in the brain. Therefore, CBP not only has the potential of complementing or replacing BP in deep learning practice, but also provides a new way for understanding the learning process of the real brain.
Collapse
Affiliation(s)
- Peng Tao
- Key Laboratory of Systems Health Science of Zhejiang Province, School of Life Science, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Chinese Academy of Sciences, Hangzhou 310024, China; Key Laboratory of Systems Biology, Shanghai Institute of Biochemistry and Cell Biology, Center for Excellence in Molecular Cell Science, Chinese Academy of Sciences, Shanghai 200031, China.
| | - Jie Cheng
- Key Laboratory of Systems Biology, Shanghai Institute of Biochemistry and Cell Biology, Center for Excellence in Molecular Cell Science, Chinese Academy of Sciences, Shanghai 200031, China.
| | - Luonan Chen
- Key Laboratory of Systems Health Science of Zhejiang Province, School of Life Science, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Chinese Academy of Sciences, Hangzhou 310024, China; Key Laboratory of Systems Biology, Shanghai Institute of Biochemistry and Cell Biology, Center for Excellence in Molecular Cell Science, Chinese Academy of Sciences, Shanghai 200031, China; Guangdong Institute of Intelligence Science and Technology, Hengqin, Zhuhai, Guangdong 519031, China.
| |
Collapse
|
15
|
Abstract
Vision and learning have long been considered to be two areas of research linked only distantly. However, recent developments in vision research have changed the conceptual definition of vision from a signal-evaluating process to a goal-oriented interpreting process, and this shift binds learning, together with the resulting internal representations, intimately to vision. In this review, we consider various types of learning (perceptual, statistical, and rule/abstract) associated with vision in the past decades and argue that they represent differently specialized versions of the fundamental learning process, which must be captured in its entirety when applied to complex visual processes. We show why the generalized version of statistical learning can provide the appropriate setup for such a unified treatment of learning in vision, what computational framework best accommodates this kind of statistical learning, and what plausible neural scheme could feasibly implement this framework. Finally, we list the challenges that the field of statistical learning faces in fulfilling the promise of being the right vehicle for advancing our understanding of vision in its entirety. Expected final online publication date for the Annual Review of Vision Science, Volume 8 is September 2022. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
Collapse
Affiliation(s)
- József Fiser
- Department of Cognitive Science, Center for Cognitive Computation, Central European University, Vienna 1100, Austria;
| | - Gábor Lengyel
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, New York 14627, USA
| |
Collapse
|
16
|
Neri P. Deep networks may capture biological behavior for shallow, but not deep, empirical characterizations. Neural Netw 2022; 152:244-266. [PMID: 35567948 DOI: 10.1016/j.neunet.2022.04.023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2021] [Revised: 04/15/2022] [Accepted: 04/20/2022] [Indexed: 11/19/2022]
Abstract
We assess whether deep convolutional networks (DCN) can account for a most fundamental property of human vision: detection/discrimination of elementary image elements (bars) at different contrast levels. The human visual process can be characterized to varying degrees of "depth," ranging from percentage of correct detection to detailed tuning and operating characteristics of the underlying perceptual mechanism. We challenge deep networks with the same stimuli/tasks used with human observers and apply equivalent characterization of the stimulus-response coupling. In general, we find that popular DCN architectures do not account for signature properties of the human process. For shallow depth of characterization, some variants of network-architecture/training-protocol produce human-like trends; however, more articulate empirical descriptors expose glaring discrepancies. Networks can be coaxed into learning those richer descriptors by shadowing a human surrogate in the form of a tailored circuit perturbed by unstructured input, thus ruling out the possibility that human-model misalignment in standard protocols may be attributable to insufficient representational power. These results urge caution in assessing whether neural networks do or do not capture human behavior: ultimately, our ability to assess "success" in this area can only be as good as afforded by the depth of behavioral characterization against which the network is evaluated. We propose a novel set of metrics/protocols that impose stringent constraints on the evaluation of DCN behavior as an adequate approximation to biological processes.
Collapse
Affiliation(s)
- Peter Neri
- Laboratoire des Systèmes Perceptifs (UMR8248), École normale supérieure, PSL Research University, Paris, France.
| |
Collapse
|
17
|
|
18
|
Zhou L, Yang A, Meng M, Zhou K. Emerged human-like facial expression representation in a deep convolutional neural network. SCIENCE ADVANCES 2022; 8:eabj4383. [PMID: 35319988 PMCID: PMC8942361 DOI: 10.1126/sciadv.abj4383] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/12/2021] [Accepted: 02/02/2022] [Indexed: 06/14/2023]
Abstract
Recent studies found that the deep convolutional neural networks (DCNNs) trained to recognize facial identities spontaneously learned features that support facial expression recognition, and vice versa. Here, we showed that the self-emerged expression-selective units in a VGG-Face trained for facial identification were tuned to distinct basic expressions and, importantly, exhibited hallmarks of human expression recognition (i.e., facial expression confusion and categorical perception). We then investigated whether the emergence of expression-selective units is attributed to either face-specific experience or domain-general processing by conducting the same analysis on a VGG-16 trained for object classification and an untrained VGG-Face without any visual experience, both having the identical architecture with the pretrained VGG-Face. Although similar expression-selective units were found in both DCNNs, they did not exhibit reliable human-like characteristics of facial expression perception. Together, these findings revealed the necessity of domain-specific visual experience of face identity for the development of facial expression perception, highlighting the contribution of nurture to form human-like facial expression perception.
Collapse
Affiliation(s)
- Liqin Zhou
- Beijing Key Laboratory of Applied Experimental Psychology, Faculty of Psychology, Beijing Normal University, Beijing 100875, China
| | - Anmin Yang
- Beijing Key Laboratory of Applied Experimental Psychology, Faculty of Psychology, Beijing Normal University, Beijing 100875, China
| | - Ming Meng
- Philosophy and Social Science Laboratory of Reading and Development in Children and Adolescents (South China Normal University), Ministry of Education, Guangzhou 510631, China
- Guangdong Key Laboratory of Mental Health and Cognitive Science, School of Psychology, South China Normal University, Guangzhou 510631, China
| | - Ke Zhou
- Beijing Key Laboratory of Applied Experimental Psychology, Faculty of Psychology, Beijing Normal University, Beijing 100875, China
| |
Collapse
|
19
|
Hulse SV, Renoult JP, Mendelson TC. Using deep neural networks to model similarity between visual patterns: Application to fish sexual signals. ECOL INFORM 2022. [DOI: 10.1016/j.ecoinf.2021.101486] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
|
20
|
Aguilar-Arguello S, Taylor AH, Nelson XJ. Jumping spiders do not seem fooled by texture gradient illusions. Behav Processes 2022; 196:104603. [DOI: 10.1016/j.beproc.2022.104603] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2021] [Revised: 01/10/2022] [Accepted: 02/02/2022] [Indexed: 11/02/2022]
|
21
|
Jung H, Wager TD, Carter RM. Novel Cognitive Functions Arise at the Convergence of Macroscale Gradients. J Cogn Neurosci 2021; 34:381-396. [PMID: 34942643 DOI: 10.1162/jocn_a_01803] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Functions in higher-order brain regions are the source of extensive debate. Although past trends have been to describe the brain-especially posterior cortical areas-in terms of a set of functional modules, a new emerging paradigm focuses on the integration of proximal functions. In this review, we synthesize emerging evidence that a variety of novel functions in the higher-order brain regions are due to convergence: convergence of macroscale gradients brings feature-rich representations into close proximity, presenting an opportunity for novel functions to arise. Using the TPJ as an example, we demonstrate that convergence is enabled via three properties of the brain: (1) hierarchical organization, (2) abstraction, and (3) equidistance. As gradients travel from primary sensory cortices to higher-order brain regions, information becomes abstracted and hierarchical, and eventually, gradients meet at a point maximally and equally distant from their sensory origins. This convergence, which produces multifaceted combinations, such as mentalizing another person's thought or projecting into a future space, parallels evolutionary and developmental characteristics in such regions, resulting in new cognitive and affective faculties.
Collapse
Affiliation(s)
- Heejung Jung
- University of Colorado Boulder.,Dartmouth College
| | - Tor D Wager
- University of Colorado Boulder.,Dartmouth College
| | | |
Collapse
|
22
|
Herpers J, Arsenault JT, Vanduffel W, Vogels R. Stimulation of the ventral tegmental area induces visual cortical plasticity at the neuronal level. Cell Rep 2021; 37:109998. [PMID: 34758325 DOI: 10.1016/j.celrep.2021.109998] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Revised: 09/20/2021] [Accepted: 10/22/2021] [Indexed: 11/17/2022] Open
Abstract
fMRI studies have shown that pairing a task-irrelevant visual feature with electrical micro-stimulation of the ventral tegmental area (VTA-EM) is sufficient to increase the sensory cortical representation of the paired feature and to improve perceptual performance. However, since fMRI provides an indirect measure of neural activity, the neural response changes underlying the fMRI activations are unknown. Here, we pair a task-irrelevant grating orientation with VTA-EM while attention is directed to a difficult orthogonal task. We examine the changes in neural response properties in macaques by recording spiking activity in the posterior inferior temporal cortex, the locus of fMRI-defined plasticity in previous studies. We observe a relative increase in mean spike rate and preference for the VTA-EM paired orientation compared to an unpaired orientation, which is unrelated to attention. These results demonstrate that VTA-EM-stimulus pairing is sufficient to induce sensory cortical plasticity at the spiking level in nonhuman primates.
Collapse
Affiliation(s)
- Jerome Herpers
- Laboratory for Neuro- and Psychophysiology, Department of Neurosciences, KU Leuven Medical School, 3000 Leuven, Belgium; Leuven Brain Institute, KU Leuven, 3000 Leuven, Belgium
| | - John T Arsenault
- Laboratory for Neuro- and Psychophysiology, Department of Neurosciences, KU Leuven Medical School, 3000 Leuven, Belgium; Leuven Brain Institute, KU Leuven, 3000 Leuven, Belgium
| | - Wim Vanduffel
- Laboratory for Neuro- and Psychophysiology, Department of Neurosciences, KU Leuven Medical School, 3000 Leuven, Belgium; Leuven Brain Institute, KU Leuven, 3000 Leuven, Belgium; Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA 02129, USA; Department of Radiology, Harvard Medical School, Boston, MA 02144, USA
| | - Rufin Vogels
- Laboratory for Neuro- and Psychophysiology, Department of Neurosciences, KU Leuven Medical School, 3000 Leuven, Belgium; Leuven Brain Institute, KU Leuven, 3000 Leuven, Belgium.
| |
Collapse
|
23
|
Liu YH, Zhu J, Constantinidis C, Zhou X. Emergence of prefrontal neuron maturation properties by training recurrent neural networks in cognitive tasks. iScience 2021; 24:103178. [PMID: 34667944 PMCID: PMC8506971 DOI: 10.1016/j.isci.2021.103178] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2021] [Revised: 06/16/2021] [Accepted: 09/22/2021] [Indexed: 01/14/2023] Open
Abstract
Working memory and response inhibition are functions that mature relatively late in life, after adolescence, paralleling the maturation of the prefrontal cortex. The link between behavioral and neural maturation is not obvious, however, making it challenging to understand how neural activity underlies the maturation of cognitive function. To gain insights into the nature of observed changes in prefrontal activity between adolescence and adulthood, we investigated the progressive changes in unit activity of recurrent neural networks as they were trained to perform working memory and response inhibition tasks. These included increased delay period activity during working memory tasks and increased activation in antisaccade tasks. These findings reveal universal properties underlying the neuronal computations behind cognitive tasks and explicate the nature of changes that occur as the result of developmental maturation. Properties of RNN networks during training offer insights in prefrontal maturation Fully trained networks exhibit higher levels of activity in working memory tasks Trained networks also exhibit higher activation in antisaccade tasks Partially trained RNNs can generate accurate predictions of immature PFC activity
Collapse
Affiliation(s)
- Yichen Henry Liu
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, USA
| | - Junda Zhu
- Neuroscience Program, Vanderbilt University, Nashville, TN 37235, USA
| | - Christos Constantinidis
- Neuroscience Program, Vanderbilt University, Nashville, TN 37235, USA.,Department of Biomedical Engineering, Vanderbilt University, Nashville, TN 37235, USA.,Department of Ophthalmology and Visual Sciences, Vanderbilt University Medical Center, Nashville, TN 37232, USA
| | - Xin Zhou
- Department of Computer Science, Vanderbilt University, Nashville, TN 37235, USA.,Department of Biomedical Engineering, Vanderbilt University, Nashville, TN 37235, USA.,Data Science Institute, Vanderbilt University, Nashville, TN 37235, USA
| |
Collapse
|
24
|
Sun ED, Dekel R. ImageNet-trained deep neural networks exhibit illusion-like response to the Scintillating grid. J Vis 2021; 21:15. [PMID: 34677575 PMCID: PMC8543405 DOI: 10.1167/jov.21.11.15] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Deep neural network (DNN) models for computer vision are capable of human-level object recognition. Consequently, similarities between DNN and human vision are of interest. Here, we characterize DNN representations of Scintillating grid visual illusion images in which white disks are perceived to be partially black. Specifically, we use VGG-19 and ResNet-101 DNN models that were trained for image classification and consider the representational dissimilarity (\(L^1\) distance in the penultimate layer) between pairs of images: one with white Scintillating grid disks and the other with disks of decreasing luminance levels. Results showed a nonmonotonic relation, such that decreasing disk luminance led to an increase and subsequently a decrease in representational dissimilarity. That is, the Scintillating grid image with white disks was closer, in terms of the representation, to images with black disks than images with gray disks. In control nonillusion images, such nonmonotonicity was rare. These results suggest that nonmonotonicity in a deep computational representation is a potential test for illusion-like response geometry in DNN models.
Collapse
Affiliation(s)
- Eric D Sun
- Mather House, Harvard University, Cambridge, MA, USA.,
| | - Ron Dekel
- Department of Neurobiology, Weizmann Institute of Science, Rehovot, PA, Israel.,
| |
Collapse
|
25
|
Lindsay GW. Convolutional Neural Networks as a Model of the Visual System: Past, Present, and Future. J Cogn Neurosci 2021; 33:2017-2031. [DOI: 10.1162/jocn_a_01544] [Citation(s) in RCA: 96] [Impact Index Per Article: 32.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Abstract
Abstract
Convolutional neural networks (CNNs) were inspired by early findings in the study of biological vision. They have since become successful tools in computer vision and state-of-the-art models of both neural activity and behavior on visual tasks. This review highlights what, in the context of CNNs, it means to be a good model in computational neuroscience and the various ways models can provide insight. Specifically, it covers the origins of CNNs and the methods by which we validate them as models of biological vision. It then goes on to elaborate on what we can learn about biological vision by understanding and experimenting on CNNs and discusses emerging opportunities for the use of CNNs in vision research beyond basic object recognition.
Collapse
|
26
|
Abstract
Perceptual learning has been widely used to study the plasticity of the visual system in adults. Owing to the belief that practice makes perfect, perceptual learning protocols usually require subjects to practice a task thousands of times over days, even weeks. However, we know very little about the relationship between training amount and behavioral improvement. Here, four groups of subjects underwent motion direction discrimination training over 8 days with 40, 120, 360, or 1080 trials per day. Surprisingly, different daily training amounts induced similar improvement across the four groups, and the similarity lasted for at least 2 weeks. Moreover, the group with 40 training trials per day showed more learning transfer from the trained direction to the untrained directions than the group with 1080 training trials per day immediately after training and 2 weeks later. These findings suggest that perceptual learning of motion direction discrimination is not always dependent on the daily training amount and less training leads to more transfer.
Collapse
Affiliation(s)
- Yongqian Song
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, People's Republic of China.,IDG/McGovern Institute for Brain Research, Peking University, Beijing, People's Republic of China.,Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, People's Republic of China.,
| | - Nihong Chen
- Department of Psychology, Tsinghua University, Beijing, People's Republic of China.,IDG/McGovern Institute for Brain Research, Tsinghua University, Beijing, People's Republic of China.,
| | - Fang Fang
- School of Psychological and Cognitive Sciences and Beijing Key Laboratory of Behavior and Mental Health, Peking University, Beijing, People's Republic of China.,IDG/McGovern Institute for Brain Research, Peking University, Beijing, People's Republic of China.,Peking-Tsinghua Center for Life Sciences, Peking University, Beijing, People's Republic of China.,
| |
Collapse
|
27
|
Saxe A, Nelli S, Summerfield C. If deep learning is the answer, what is the question? Nat Rev Neurosci 2020; 22:55-67. [PMID: 33199854 DOI: 10.1038/s41583-020-00395-8] [Citation(s) in RCA: 119] [Impact Index Per Article: 29.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/02/2020] [Indexed: 11/09/2022]
Abstract
Neuroscience research is undergoing a minor revolution. Recent advances in machine learning and artificial intelligence research have opened up new ways of thinking about neural computation. Many researchers are excited by the possibility that deep neural networks may offer theories of perception, cognition and action for biological brains. This approach has the potential to radically reshape our approach to understanding neural systems, because the computations performed by deep networks are learned from experience, and not endowed by the researcher. If so, how can neuroscientists use deep networks to model and understand biological brains? What is the outlook for neuroscientists who seek to characterize computations or neural codes, or who wish to understand perception, attention, memory and executive functions? In this Perspective, our goal is to offer a road map for systems neuroscience research in the age of deep learning. We discuss the conceptual and methodological challenges of comparing behaviour, learning dynamics and neural representations in artificial and biological systems, and we highlight new research questions that have emerged for neuroscience as a direct consequence of recent advances in machine learning.
Collapse
Affiliation(s)
- Andrew Saxe
- Department of Experimental Psychology, University of Oxford, Oxford, UK.
| | - Stephanie Nelli
- Department of Experimental Psychology, University of Oxford, Oxford, UK.
| | | |
Collapse
|
28
|
Seitz AR. Perceptual Learning: How Does the Visual Circuit Change through Experience? Curr Biol 2020; 30:R1309-R1311. [DOI: 10.1016/j.cub.2020.08.097] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
|
29
|
Frank SM, Qi A, Ravasio D, Sasaki Y, Rosen EL, Watanabe T. Supervised Learning Occurs in Visual Perceptual Learning of Complex Natural Images. Curr Biol 2020; 30:2995-3000.e3. [PMID: 32502415 DOI: 10.1016/j.cub.2020.05.050] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2020] [Revised: 04/14/2020] [Accepted: 05/14/2020] [Indexed: 01/13/2023]
Abstract
There have been long-standing debates regarding whether supervised or unsupervised learning mechanisms are involved in visual perceptual learning (VPL) [1-14]. However, these debates have been based on the effects of simple feedback only about response accuracy in detection or discrimination tasks of low-level visual features such as orientation [15-22]. Here, we examined whether the content of response feedback plays a critical role for the acquisition and long-term retention of VPL of complex natural images. We trained three groups of human subjects (n = 72 in total) to better detect "grouped microcalcifications" or "architectural distortion" lesions (referred to as calcification and distortion in the following) in mammograms either with no trial-by-trial feedback, partial trial-by-trial feedback (response correctness only), or detailed trial-by-trial feedback (response correctness and target location). Distortion lesions consist of more complex visual structures than calcification lesions [23-26]. We found that partial feedback is necessary for VPL of calcifications, whereas detailed feedback is required for VPL of distortions. Furthermore, detailed feedback during training is necessary for VPL of distortion and calcification lesions to be retained for 6 months. These results show that although supervised learning is heavily involved in VPL of complex natural images, the extent of supervision for VPL varies across different types of complex natural images. Such differential requirements for VPL to improve the detectability of lesions in mammograms are potentially informative for the professional training of radiologists.
Collapse
Affiliation(s)
- Sebastian M Frank
- Brown University, Department of Cognitive, Linguistic, and Psychological Sciences, 190 Thayer Street, Providence, RI 02912, USA.
| | - Andrea Qi
- Brown University, Department of Cognitive, Linguistic, and Psychological Sciences, 190 Thayer Street, Providence, RI 02912, USA
| | - Daniela Ravasio
- Brown University, Department of Cognitive, Linguistic, and Psychological Sciences, 190 Thayer Street, Providence, RI 02912, USA
| | - Yuka Sasaki
- Brown University, Department of Cognitive, Linguistic, and Psychological Sciences, 190 Thayer Street, Providence, RI 02912, USA
| | - Eric L Rosen
- Stanford University, Department of Radiology, 300 Pasteur Drive, Stanford, CA 94305, USA; University of Colorado Denver, Department of Radiology, 12401 East 17th Avenue, Aurora, CO 80045, USA
| | - Takeo Watanabe
- Brown University, Department of Cognitive, Linguistic, and Psychological Sciences, 190 Thayer Street, Providence, RI 02912, USA.
| |
Collapse
|
30
|
Emergent Virtual Analytics: Modeling Contextual Control of Derived Stimulus Relations. BEHAVIOR AND SOCIAL ISSUES 2020. [DOI: 10.1007/s42822-020-00032-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2023]
|
31
|
Sandeep S, Shelton CR, Pahor A, Jaeggi SM, Seitz AR. Application of Machine Learning Models for Tracking Participant Skills in Cognitive Training. Front Psychol 2020; 11:1532. [PMID: 32793032 PMCID: PMC7387708 DOI: 10.3389/fpsyg.2020.01532] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2019] [Accepted: 06/09/2020] [Indexed: 11/13/2022] Open
Abstract
A key need in cognitive training interventions is to personalize task difficulty to each user and to adapt this difficulty to continually apply appropriate challenges as users improve their skill to perform the tasks. Here we examine how Bayesian filtering approaches, such as hidden Markov models and Kalman filters, and deep-learning approaches, such as the long short-term memory (LSTM) model, may be useful methods to estimate user skill level and predict appropriate task challenges. A possible advantage of these models over commonly used adaptive methods, such as staircases or blockwise adjustment methods that are based only upon recent performance, is that Bayesian filtering and deep learning approaches can model the trajectory of user performance across multiple sessions and incorporate data from multiple users to optimize local estimates. As a proof of concept, we fit data from two large cohorts of undergraduate students performing WM training using an N-back task. Results show that all three models predict appropriate challenges for different users. However, the hidden Markov models were most accurate in predicting participants' performances as a function of provided challenges, and thus, they placed participants at appropriate future challenges. These data provide good support for the potential of machine learning approaches as appropriate methods to personalize task performance to users in tasks that require adaptively determined challenges.
Collapse
Affiliation(s)
- Sanjana Sandeep
- Department of Computer Science, University of California, Riverside, Riverside, CA, United States
- Brain Game Center, University of California, Riverside, Riverside, CA, United States
| | - Christian R. Shelton
- Department of Computer Science, University of California, Riverside, Riverside, CA, United States
| | - Anja Pahor
- Brain Game Center, University of California, Riverside, Riverside, CA, United States
- School of Education, University of California, Irvine, Irvine, CA, United States
| | - Susanne M. Jaeggi
- School of Education, University of California, Irvine, Irvine, CA, United States
| | - Aaron R. Seitz
- Brain Game Center, University of California, Riverside, Riverside, CA, United States
- Department of Psychology, University of California, Riverside, Riverside, CA, United States
| |
Collapse
|
32
|
Le Dantec CC, Seitz AR. Dissociating electrophysiological correlates of contextual and perceptual learning in a visual search task. J Vis 2020; 20:7. [PMID: 32525986 PMCID: PMC7416887 DOI: 10.1167/jov.20.6.7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Perceptual learning and contextual learning are two types of implicit visual learning that can co-occur in the same tasks. For example, to find an animal in the woods, you need to know where to look in the environment (contextual learning) and you must be able to discriminate its features (perceptual learning). However, contextual and perceptual learning are typically studied using distinct experimental paradigms, and little is known regarding their comparative neural mechanisms. In this study, we investigated contextual and perceptual learning in 12 healthy adult humans as they performed the same visual search task, and we examined psychophysical and electrophysiological (event-related potentials) measures of learning. Participants were trained to look for a visual stimulus, a small line with a specific orientation, presented among distractors. We found better performance for the trained target orientation as compared to an untrained control orientation, reflecting specificity of perceptual learning for the orientation of trained elements. This orientation specificity effect was associated with changes in the C1 component. We also found better performance for repeated spatial configurations as compared to novel ones, reflecting contextual learning. This context-specific effect was associated with the N2pc component. Taken together, these results suggest that contextual and perceptual learning are distinct visual learning phenomena that have different behavioral and electrophysiological characteristics.
Collapse
|
33
|
|
34
|
Jacques T, Seitz AR. Moderating effects of visual attention and action video game play on perceptual learning with the texture discrimination task. Vision Res 2020; 171:64-72. [PMID: 32172941 DOI: 10.1016/j.visres.2020.02.003] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2017] [Revised: 02/10/2020] [Accepted: 02/11/2020] [Indexed: 10/24/2022]
Abstract
There is currently substantial controversy regarding the reliability of observed patterns of perceptual learning. Contributing to this controversy are a lack of accounting for individual differences and how variations in training can give rise to different patterns of learning. Here we sought to investigate the impact of individual differences in attention, as measured with the Useful Field of View (UFOV) task, and action video game use on perceptual learning in a large sample of subjects trained on a Texture Discrimination Task (TDT). We examined baseline performance on the TDT, learning on the initially trained TDT stimuli and transfer to a subsequently trained background orientation. We find that participants showing better performance on the UFOV task performed better on the TDT, and also showed greater learning and transfer to an untrained background orientation. On the other hand, self-report of action video game play only inconsistently related performance, learning or transfer on the TDT. Further, we failed to replicate previous findings that training with different backgrounds gives rise to interference on the TDT. Together these results suggest that, while differences between individuals and differences in task structure play a role in perceptual learning, previous findings on the impact of action video game use and interference between training stimuli in perceptual learning may be idiosyncratic.
Collapse
|
35
|
German JS, Jacobs RA. Can machine learning account for human visual object shape similarity judgments? Vision Res 2020; 167:87-99. [PMID: 31972448 DOI: 10.1016/j.visres.2019.12.001] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2019] [Revised: 10/22/2019] [Accepted: 12/12/2019] [Indexed: 11/27/2022]
Abstract
We describe and analyze the performance of metric learning systems, including deep neural networks (DNNs), on a new dataset of human visual object shape similarity judgments of naturalistic, part-based objects known as "Fribbles". In contrast to previous studies which asked participants to judge similarity when objects or scenes were rendered from a single viewpoint, we rendered Fribbles from multiple viewpoints and asked participants to judge shape similarity in a viewpoint-invariant manner. Metrics trained using pixel-based or DNN-based representations fail to explain our experimental data, but a metric trained with a viewpoint-invariant, part-based representation produces a good fit. We also find that although neural networks can learn to extract the part-based representation-and therefore should be capable of learning to model our data-networks trained with a "triplet loss" function based on similarity judgments do not perform well. We analyze this failure, providing a mathematical description of the relationship between the metric learning objective function and the triplet loss function. The poor performance of neural networks appears to be due to the nonconvexity of the optimization problem in network weight space. We conclude that viewpoint insensitivity is a critical aspect of human visual shape perception, and that neural network and other machine learning methods will need to learn viewpoint-insensitive representations in order to account for people's visual object shape similarity judgments.
Collapse
Affiliation(s)
- Joseph Scott German
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, NY 14627, United States.
| | - Robert A Jacobs
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, NY 14627, United States.
| |
Collapse
|
36
|
Song Y, Lukasiewicz T, Xu Z, Bogacz R. Can the Brain Do Backpropagation? -Exact Implementation of Backpropagation in Predictive Coding Networks. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 2020; 33:22566-22579. [PMID: 33840988 PMCID: PMC7610561] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Backpropagation (BP) has been the most successful algorithm used to train artificial neural networks. However, there are several gaps between BP and learning in biologically plausible neuronal networks of the brain (learning in the brain, or simply BL, for short), in particular, (1) it has been unclear to date, if BP can be implemented exactly via BL, (2) there is a lack of local plasticity in BP, i.e., weight updates require information that is not locally available, while BL utilizes only locally available information, and (3) there is a lack of autonomy in BP, i.e., some external control over the neural network is required (e.g., switching between prediction and learning stages requires changes to dynamics and synaptic plasticity rules), while BL works fully autonomously. Bridging such gaps, i.e., understanding how BP can be approximated by BL, has been of major interest in both neuroscience and machine learning. Despite tremendous efforts, however, no previous model has bridged the gaps at a degree of demonstrating an equivalence to BP, instead, only approximations to BP have been shown. Here, we present for the first time a framework within BL that bridges the above crucial gaps. We propose a BL model that (1) produces exactly the same updates of the neural weights as BP, while (2) employing local plasticity, i.e., all neurons perform only local computations, done simultaneously. We then modify it to an alternative BL model that (3) also works fully autonomously. Overall, our work provides important evidence for the debate on the long-disputed question whether the brain can perform BP.
Collapse
Affiliation(s)
- Yuhang Song
- Department of Computer Science, University of Oxford, UK
| | | | - Zhenghua Xu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, China
| | - Rafal Bogacz
- MRC Brain Network Dynamics Unit, University of Oxford, UK
| |
Collapse
|
37
|
|
38
|
Wandeto JM, Dresp-Langley B. The quantization error in a Self-Organizing Map as a contrast and colour specific indicator of single-pixel change in large random patterns. Neural Netw 2019; 120:116-128. [PMID: 31610898 DOI: 10.1016/j.neunet.2019.09.017] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The quantization error in a fixed-size Self-Organizing Map (SOM) with unsupervised winner-take-all learning has previously been used successfully to detect, in minimal computation time, highly meaningful changes across images in medical time series and in time series of satellite images. Here, the functional properties of the quantization error in SOM are explored further to show that the metric is capable of reliably discriminating between the finest differences in local contrast intensities and contrast signs. While this capability of the QE is akin to functional characteristics of a specific class of retinal ganglion cells (the so-called Y-cells) in the visual systems of the primate and the cat, the sensitivity of the QE surpasses the capacity limits of human visual detection. Here, the quantization error in the SOM is found to reliably signal changes in contrast or colour when contrast information is removed from or added to the image, but not when the amount and relative weight of contrast information is constant and only the local spatial position of contrast elements in the pattern changes. While the RGB Mean reflects coarser changes in colour or contrast well enough, the SOM-QE is shown to outperform the RGB Mean in the detection of single-pixel changes in images with up to five million pixels. This could have important implications in the context of unsupervised image learning and computational building block approaches to large sets of image data (big data), including deep learning blocks, and automatic detection of contrast change at the nanoscale in Transmission or Scanning Electron Micrographs (TEM, SEM), or at the subpixel level in multispectral and hyper-spectral imaging data.
Collapse
Affiliation(s)
- John M Wandeto
- Dedan Kimathi University of Technology, Department of Information Technology, Nyeri, Kenya
| | - Birgitta Dresp-Langley
- Centre National de la Recherche Scientifique (CNRS), UMR 7357 ICube Lab, University of Strasbourg, France.
| |
Collapse
|
39
|
Interfering with a memory without erasing its trace. Neural Netw 2019; 121:339-355. [PMID: 31593840 DOI: 10.1016/j.neunet.2019.09.027] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2019] [Revised: 09/16/2019] [Accepted: 09/22/2019] [Indexed: 11/21/2022]
Abstract
Previous research has shown that performance of a novice skill can be easily interfered with by subsequent training of another skill. We address the open questions whether extensively trained skills show the same vulnerability to interference as novice skills and which memory mechanism regulates interference between expert skills. We developed a recurrent neural network model of V1 able to learn from feedback experienced over the course of a long-term orientation discrimination experiment. After first exposing the model to one discrimination task for 3480 consecutive trials, we assessed how its performance was affected by subsequent training in a second, similar task. Training the second task strongly interfered with the first (highly trained) discrimination skill. The magnitude of interference depended on the relative amounts of training devoted to the different tasks. We used these and other model outcomes as predictions for a perceptual learning experiment in which human participants underwent the same training protocol as our model. Specifically, over the course of three months participants underwent baseline training in one orientation discrimination task for 15 sessions before being trained for 15 sessions on a similar task and finally undergoing another 15 sessions of training on the first task (to assess interference). Across all conditions, the pattern of interference observed empirically closely matched model predictions. According to our model, behavioral interference can be explained by antagonistic changes in neuronal tuning induced by the two tasks. Remarkably, this did not stem from erasing connections due to earlier learning but rather from a reweighting of lateral inhibition.
Collapse
|
40
|
Delius JD, Delius JAM. Systematic Analysis of Pigeons' Discrimination of Pixelated Stimuli: A Hierarchical Pattern Recognition System Is Not Identifiable. Sci Rep 2019; 9:13929. [PMID: 31558750 PMCID: PMC6763494 DOI: 10.1038/s41598-019-50212-1] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2019] [Accepted: 09/06/2019] [Indexed: 02/07/2023] Open
Abstract
Pigeons learned to discriminate two different patterns displayed with miniature light-emitting diode arrays. They were then tested with 84 interspersed, non-reinforced degraded pattern pairs. Choices ranged between 100% and 50% for one or other of the patterns. Stimuli consisting of few pixels yielded low choice scores whereas those consisting of many pixels yielded a broad range of scores. Those patterns with a high number of pixels coinciding with those of the rewarded training stimulus were preferred and those with a high number of pixels coinciding with the non-rewarded training pattern were avoided; a discrimination index based on this correlated 0.74 with the pattern choices. Pixels common to both training patterns had a minimal influence. A pixel-by-pixel analysis revealed that eight pixels of one pattern and six pixels of the other pattern played a prominent role in the pigeons’ choices. These pixels were disposed in four and two clusters of neighbouring locations. A summary index calculated on this basis still only yielded a weak 0.73 correlation. The individual pigeons’ data furthermore showed that these clusters were a mere averaging mirage. The pigeons’ performance depends on deep learning in a midbrain-based multimillion synapse neuronal network. Pixelated visual patterns should be helpful when simulating perception of patterns with artificial networks.
Collapse
Affiliation(s)
- Juan D Delius
- Experimental Psychology, University of Konstanz, 78457, Konstanz, Germany.
| | - Julia A M Delius
- Center for Lifespan Psychology, Max Planck Institute for Human Development, 14195, Berlin, Germany
| |
Collapse
|
41
|
Jacobs RA, Xu C. Can multisensory training aid visual learning? A computational investigation. J Vis 2019; 19:1. [PMID: 31480074 DOI: 10.1167/19.11.1] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Although real-world environments are often multisensory, visual scientists typically study visual learning in unisensory environments containing visual signals only. Here, we use deep or artificial neural networks to address the question, Can multisensory training aid visual learning? We examine a network's internal representations of objects based on visual signals in two conditions: (a) when the network is initially trained with both visual and haptic signals, and (b) when it is initially trained with visual signals only. Our results demonstrate that a network trained in a visual-haptic environment (in which visual, but not haptic, signals are orientation-dependent) tends to learn visual representations containing useful abstractions, such as the categorical structure of objects, and also learns representations that are less sensitive to imaging parameters, such as viewpoint or orientation, that are irrelevant for object recognition or classification tasks. We conclude that researchers studying perceptual learning in vision-only contexts may be overestimating the difficulties associated with important perceptual learning problems. Although multisensory perception has its own challenges, perceptual learning can become easier when it is considered in a multisensory setting.
Collapse
Affiliation(s)
- Robert A Jacobs
- Department of Brain and Cognitive Sciences, University of Rochester, Rochester, NY, USA
| | - Chenliang Xu
- Department of Computer Science, University of Rochester, Rochester, NY, USA
| |
Collapse
|
42
|
Wandeto JM, Dresp-Langley B. The quantization error in a Self-Organizing Map as a contrast and colour specific indicator of single-pixel change in large random patterns. Neural Netw 2019; 119:273-285. [PMID: 31473578 DOI: 10.1016/j.neunet.2019.08.014] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2019] [Revised: 07/08/2019] [Accepted: 08/09/2019] [Indexed: 10/26/2022]
Abstract
The quantization error in a fixed-size Self-Organizing Map (SOM) with unsupervised winner-take-all learning has previously been used successfully to detect, in minimal computation time, highly meaningful changes across images in medical time series and in time series of satellite images. Here, the functional properties of the quantization error in SOM are explored further to show that the metric is capable of reliably discriminating between the finest differences in local contrast intensities and contrast signs. While this capability of the QE is akin to functional characteristics of a specific class of retinal ganglion cells (the so-called Y-cells) in the visual systems of the primate and the cat, the sensitivity of the QE surpasses the capacity limits of human visual detection. Here, the quantization error in the SOM is found to reliably signal changes in contrast or colour when contrast information is removed from or added to the image, but not when the amount and relative weight of contrast information is constant and only the local spatial position of contrast elements in the pattern changes. While the RGB Mean reflects coarser changes in colour or contrast well enough, the SOM-QE is shown to outperform the RGB Mean in the detection of single-pixel changes in images with up to five million pixels. This could have important implications in the context of unsupervised image learning and computational building block approaches to large sets of image data (big data), including deep learning blocks, and automatic detection of contrast change at the nanoscale in Transmission or Scanning Electron Micrographs (TEM, SEM), or at the subpixel level in multispectral and hyper-spectral imaging data.
Collapse
Affiliation(s)
- John M Wandeto
- Dedan Kimathi University of Technology, Department of Information Technology, Nyeri, Kenya
| | - Birgitta Dresp-Langley
- Centre National de la Recherche Scientifique (CNRS), UMR 7357 ICube Lab, University of Strasbourg, France.
| |
Collapse
|
43
|
Occam’s Razor for Big Data? On Detecting Quality in Large Unstructured Datasets. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9153065] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/05/2023]
Abstract
Detecting quality in large unstructured datasets requires capacities far beyond the limits of human perception and communicability and, as a result, there is an emerging trend towards increasingly complex analytic solutions in data science to cope with this problem. This new trend towards analytic complexity represents a severe challenge for the principle of parsimony (Occam’s razor) in science. This review article combines insight from various domains such as physics, computational science, data engineering, and cognitive science to review the specific properties of big data. Problems for detecting data quality without losing the principle of parsimony are then highlighted on the basis of specific examples. Computational building block approaches for data clustering can help to deal with large unstructured datasets in minimized computation time, and meaning can be extracted rapidly from large sets of unstructured image or video data parsimoniously through relatively simple unsupervised machine learning algorithms. Why we still massively lack in expertise for exploiting big data wisely to extract relevant information for specific tasks, recognize patterns and generate new information, or simply store and further process large amounts of sensor data is then reviewed, and examples illustrating why we need subjective views and pragmatic methods to analyze big data contents are brought forward. The review concludes on how cultural differences between East and West are likely to affect the course of big data analytics, and the development of increasingly autonomous artificial intelligence (AI) aimed at coping with the big data deluge in the near future.
Collapse
|
44
|
Deep neural network models of sensory systems: windows onto the role of task constraints. Curr Opin Neurobiol 2019; 55:121-132. [DOI: 10.1016/j.conb.2019.02.003] [Citation(s) in RCA: 51] [Impact Index Per Article: 10.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2018] [Revised: 01/13/2019] [Accepted: 02/07/2019] [Indexed: 01/05/2023]
|
45
|
Maniglia M, Seitz AR. A New Look at Visual System Plasticity. Trends Cogn Sci 2019; 23:82-83. [DOI: 10.1016/j.tics.2018.11.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2018] [Accepted: 11/08/2018] [Indexed: 11/29/2022]
|
46
|
Can Deep Learning Model Perceptual Learning? J Neurosci 2019; 39:194-196. [PMID: 30626723 DOI: 10.1523/jneurosci.2209-18.2018] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2018] [Revised: 10/30/2018] [Accepted: 11/01/2018] [Indexed: 11/21/2022] Open
|