1
|
Christensen S, Kallsen J. Is Learning in Biological Neural Networks Based on Stochastic Gradient Descent? An Analysis Using Stochastic Processes. Neural Comput 2024; 36:1424-1432. [PMID: 38669690 DOI: 10.1162/neco_a_01668] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Accepted: 02/09/2024] [Indexed: 04/28/2024]
Abstract
In recent years, there has been an intense debate about how learning in biological neural networks (BNNs) differs from learning in artificial neural networks. It is often argued that the updating of connections in the brain relies only on local information, and therefore a stochastic gradient-descent type optimization method cannot be used. In this note, we study a stochastic model for supervised learning in BNNs. We show that a (continuous) gradient step occurs approximately when each learning opportunity is processed by many local updates. This result suggests that stochastic gradient descent may indeed play a role in optimizing BNNs.
Collapse
Affiliation(s)
| | - Jan Kallsen
- Department of Mathematics, Kiel Universiy, 24118 Kiel, Germany
| |
Collapse
|
2
|
Bredenberg C, Savin C. Desiderata for Normative Models of Synaptic Plasticity. Neural Comput 2024; 36:1245-1285. [PMID: 38776950 DOI: 10.1162/neco_a_01671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 02/06/2024] [Indexed: 05/25/2024]
Abstract
Normative models of synaptic plasticity use computational rationales to arrive at predictions of behavioral and network-level adaptive phenomena. In recent years, there has been an explosion of theoretical work in this realm, but experimental confirmation remains limited. In this review, we organize work on normative plasticity models in terms of a set of desiderata that, when satisfied, are designed to ensure that a given model demonstrates a clear link between plasticity and adaptive behavior, is consistent with known biological evidence about neural plasticity and yields specific testable predictions. As a prototype, we include a detailed analysis of the REINFORCE algorithm. We also discuss how new models have begun to improve on the identified criteria and suggest avenues for further development. Overall, we provide a conceptual guide to help develop neural learning theories that are precise, powerful, and experimentally testable.
Collapse
Affiliation(s)
- Colin Bredenberg
- Center for Neural Science, New York University, New York, NY 10003, U.S.A
- Mila-Quebec AI Institute, Montréal, QC H2S 3H1, Canada
| | - Cristina Savin
- Center for Neural Science, New York University, New York, NY 10003, U.S.A
- Center for Data Science, New York University, New York, NY 10011, U.S.A.
| |
Collapse
|
3
|
Cusack R, Ranzato M, Charvet CJ. Helpless infants are learning a foundation model. Trends Cogn Sci 2024:S1364-6613(24)00114-1. [PMID: 38839537 DOI: 10.1016/j.tics.2024.05.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2023] [Revised: 04/24/2024] [Accepted: 05/03/2024] [Indexed: 06/07/2024]
Abstract
Humans have a protracted postnatal helplessness period, typically attributed to human-specific maternal constraints causing an early birth when the brain is highly immature. By aligning neurodevelopmental events across species, however, it has been found that humans are not born with especially immature brains compared with animal species with a shorter helpless period. Consistent with this, the rapidly growing field of infant neuroimaging has found that brain connectivity and functional activation at birth share many similarities with the mature brain. Inspired by machine learning, where deep neural networks also benefit from a 'helpless period' of pre-training, we propose that human infants are learning a foundation model: a set of fundamental representations that underpin later cognition with high performance and rapid generalisation.
Collapse
|
4
|
Krauhausen I, Griggs S, McCulloch I, den Toonder JMJ, Gkoupidenis P, van de Burgt Y. Bio-inspired multimodal learning with organic neuromorphic electronics for behavioral conditioning in robotics. Nat Commun 2024; 15:4765. [PMID: 38834541 DOI: 10.1038/s41467-024-48881-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 05/13/2024] [Indexed: 06/06/2024] Open
Abstract
Biological systems interact directly with the environment and learn by receiving multimodal feedback via sensory stimuli that shape the formation of internal neuronal representations. Drawing inspiration from biological concepts such as exploration and sensory processing that eventually lead to behavioral conditioning, we present a robotic system handling objects through multimodal learning. A small-scale organic neuromorphic circuit locally integrates and adaptively processes multimodal sensory stimuli, enabling the robot to interact intelligently with its surroundings. The real-time handling of sensory stimuli via low-voltage organic neuromorphic devices with synaptic functionality forms multimodal associative connections that lead to behavioral conditioning, and thus the robot learns to avoid potentially dangerous objects. This work demonstrates that adaptive neuro-inspired circuitry with multifunctional organic materials, can accommodate locally efficient bio-inspired learning for advancing intelligent robotics.
Collapse
Affiliation(s)
- Imke Krauhausen
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
- Max Planck Institute for Polymer Research, Mainz, Germany
| | - Sophie Griggs
- Department of Chemistry, University of Oxford, Oxford, UK
| | - Iain McCulloch
- Department of Chemistry, University of Oxford, Oxford, UK
| | - Jaap M J den Toonder
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | | | - Yoeri van de Burgt
- Institute for Complex Molecular Systems, Eindhoven University of Technology, Eindhoven, The Netherlands.
- Microsystems, Department of Mechanical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands.
| |
Collapse
|
5
|
Grimaldi A, Boutin V, Ieng SH, Benosman R, Perrinet LU. A robust event-driven approach to always-on object recognition. Neural Netw 2024; 178:106415. [PMID: 38852508 DOI: 10.1016/j.neunet.2024.106415] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Revised: 04/05/2024] [Accepted: 05/29/2024] [Indexed: 06/11/2024]
Abstract
We propose a neuromimetic architecture capable of always-on pattern recognition, i.e. at any time during processing. To achieve this, we have extended an existing event-based algorithm (Lagorce et al., 2017), which introduced novel spatio-temporal features as a Hierarchy Of Time-Surfaces (HOTS). Built from asynchronous events captured by a neuromorphic camera, these time surfaces allow to encode the local dynamics of a visual scene and to create an efficient event-based pattern recognition architecture. Inspired by neuroscience, we have extended this method to improve its performance. First, we add a homeostatic gain control on the activity of neurons to improve the learning of spatio-temporal patterns (Grimaldi et al., 2021). We also provide a new mathematical formalism that allows an analogy to be drawn between the HOTS algorithm and Spiking Neural Networks (SNN). Following this analogy, we transform the offline pattern categorization method into an online and event-driven layer. This classifier uses the spiking output of the network to define new time surfaces and we then perform the online classification with a neuromimetic implementation of a multinomial logistic regression. These improvements not only consistently increase the performance of the network, but also bring this event-driven pattern recognition algorithm fully online. The results have been validated on different datasets: Poker-DVS (Serrano-Gotarredona and Linares-Barranco, 2015), N-MNIST (Orchard, Jayawant et al., 2015) and DVS Gesture (Amir et al., 2017). This demonstrates the efficiency of this bio-realistic SNN for ultra-fast object recognition through an event-by-event categorization process.
Collapse
Affiliation(s)
- Antoine Grimaldi
- Aix-Marseille Universit, Institut de Neurosciences de la Timone, CNRS, Marseille, France.
| | - Victor Boutin
- Carney Institute for Brain Science, Brown University, Providence, RI, United States; Artificial and Natural Intelligence Toulouse Institute, Université de Toulouse, Toulouse, France.
| | - Sio-Hoi Ieng
- Institut de la Vision, Sorbonne Université, CNRS, Paris, France.
| | - Ryad Benosman
- Robotics Institute, Carnegie Mellon University, Pittsburg, PA, United States.
| | - Laurent U Perrinet
- Aix-Marseille Universit, Institut de Neurosciences de la Timone, CNRS, Marseille, France.
| |
Collapse
|
6
|
Jin Y, Sharifi A, Li Z, Chen S, Zeng S, Zhao S. Carbon emission prediction models: A review. THE SCIENCE OF THE TOTAL ENVIRONMENT 2024; 927:172319. [PMID: 38599410 DOI: 10.1016/j.scitotenv.2024.172319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/27/2023] [Revised: 03/26/2024] [Accepted: 04/06/2024] [Indexed: 04/12/2024]
Abstract
Amidst growing concerns over the greenhouse effect, especially its consequential impacts, establishing effective Carbon Emission Prediction Models (CEPMs) to comprehend and predict CO2 emission trends is imperative for climate change mitigation. A review of 147 Carbon Emission Prediction Model (CEPM) studies revealed three predominant functions-prediction, optimization, and prediction factor selection. Statistical models, comprising 75 instances, were the most prevalent among prediction models, followed by neural network models at 21.8 %. The consistent rise in neural network model usage, particularly feedforward architectures, was observed from 2019 to 2022. A majority of CEPMs incorporated optimized approaches, with 94.4 % utilizing metaheuristic models. Parameter optimization was the primary focus, followed by structure optimization. Prediction factor selection models, employing Grey Relational Analysis (GRA) and Principal Component Analysis (PCA) for statistical and machine learning models, respectively, filtered factors effectively. Scrutinizing accuracy, pre-optimized CEPMs exhibited varied performance, Root Mean Square Error (RMSE) values spanned from 0.112 to 1635 Mt, while post-optimization led to a notable improvement, the minimum RMSE reached 0.0003 Mt, and the maximum was 95.14 Mt. Finally, we summarized the pros and cons of existing models, classified and counted the factors that influence carbon emissions, clarified the research objectives in CEPM and assessed the applied model evaluation methods and the spatial and temporal scales of existing research.
Collapse
Affiliation(s)
- Yukai Jin
- Urban Environmental Science Lab (URBES), Graduate School of Innovation and Practice for Smart Society, Hiroshima University, Higashi-Hiroshima, 739-8529, Japan; School of Civil and Transportation Engineering, Guangdong University of Technology, Guangdong, 510006, China
| | - Ayyoob Sharifi
- The IDEC Institute, Hiroshima University, Higashi-Hiroshima, 739-8529, Japan; School of Architecture and Design, Lebanese American University, Beirut, Lebanon.
| | - Zhisheng Li
- School of Civil and Transportation Engineering, Guangdong University of Technology, Guangdong, 510006, China
| | - Sirui Chen
- School of Civil and Transportation Engineering, Guangdong University of Technology, Guangdong, 510006, China
| | - Suzhen Zeng
- School of Civil and Transportation Engineering, Guangdong University of Technology, Guangdong, 510006, China; School of Ocean Engineering and Technology, Sun Yat-sen University, Guangdong, 519000, China
| | - Shanlun Zhao
- School of Civil and Transportation Engineering, Guangdong University of Technology, Guangdong, 510006, China
| |
Collapse
|
7
|
Daruwalla K, Lipasti M. Information bottleneck-based Hebbian learning rule naturally ties working memory and synaptic updates. Front Comput Neurosci 2024; 18:1240348. [PMID: 38818385 PMCID: PMC11137249 DOI: 10.3389/fncom.2024.1240348] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 04/26/2024] [Indexed: 06/01/2024] Open
Abstract
Deep neural feedforward networks are effective models for a wide array of problems, but training and deploying such networks presents a significant energy cost. Spiking neural networks (SNNs), which are modeled after biologically realistic neurons, offer a potential solution when deployed correctly on neuromorphic computing hardware. Still, many applications train SNNs offline, and running network training directly on neuromorphic hardware is an ongoing research problem. The primary hurdle is that back-propagation, which makes training such artificial deep networks possible, is biologically implausible. Neuroscientists are uncertain about how the brain would propagate a precise error signal backward through a network of neurons. Recent progress addresses part of this question, e.g., the weight transport problem, but a complete solution remains intangible. In contrast, novel learning rules based on the information bottleneck (IB) train each layer of a network independently, circumventing the need to propagate errors across layers. Instead, propagation is implicit due the layers' feedforward connectivity. These rules take the form of a three-factor Hebbian update a global error signal modulates local synaptic updates within each layer. Unfortunately, the global signal for a given layer requires processing multiple samples concurrently, and the brain only sees a single sample at a time. We propose a new three-factor update rule where the global signal correctly captures information across samples via an auxiliary memory network. The auxiliary network can be trained a priori independently of the dataset being used with the primary network. We demonstrate comparable performance to baselines on image classification tasks. Interestingly, unlike back-propagation-like schemes where there is no link between learning and memory, our rule presents a direct connection between working memory and synaptic updates. To the best of our knowledge, this is the first rule to make this link explicit. We explore these implications in initial experiments examining the effect of memory capacity on learning performance. Moving forward, this work suggests an alternate view of learning where each layer balances memory-informed compression against task performance. This view naturally encompasses several key aspects of neural computation, including memory, efficiency, and locality.
Collapse
Affiliation(s)
- Kyle Daruwalla
- Cold Spring Harbor Laboratory, Long Island, NY, United States
| | - Mikko Lipasti
- Electrical and Computer Engineering Department, University of Wisconsin-Madison, Madison, WI, United States
| |
Collapse
|
8
|
Giallanza T, Campbell D, Cohen JD. Toward the Emergence of Intelligent Control: Episodic Generalization and Optimization. Open Mind (Camb) 2024; 8:688-722. [PMID: 38828434 PMCID: PMC11142636 DOI: 10.1162/opmi_a_00143] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2023] [Accepted: 04/01/2024] [Indexed: 06/05/2024] Open
Abstract
Human cognition is unique in its ability to perform a wide range of tasks and to learn new tasks quickly. Both abilities have long been associated with the acquisition of knowledge that can generalize across tasks and the flexible use of that knowledge to execute goal-directed behavior. We investigate how this emerges in a neural network by describing and testing the Episodic Generalization and Optimization (EGO) framework. The framework consists of an episodic memory module, which rapidly learns relationships between stimuli; a semantic pathway, which more slowly learns how stimuli map to responses; and a recurrent context module, which maintains a representation of task-relevant context information, integrates this over time, and uses it both to recall context-relevant memories (in episodic memory) and to bias processing in favor of context-relevant features and responses (in the semantic pathway). We use the framework to address empirical phenomena across reinforcement learning, event segmentation, and category learning, showing in simulations that the same set of underlying mechanisms accounts for human performance in all three domains. The results demonstrate how the components of the EGO framework can efficiently learn knowledge that can be flexibly generalized across tasks, furthering our understanding of how humans can quickly learn how to perform a wide range of tasks-a capability that is fundamental to human intelligence.
Collapse
Affiliation(s)
- Tyler Giallanza
- Department of Psychology, Princeton University, Princeton, NJ, USA
| | - Declan Campbell
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA
| | - Jonathan D. Cohen
- Department of Psychology, Princeton University, Princeton, NJ, USA
- Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA
| |
Collapse
|
9
|
Alonso N, Krichmar JL. A sparse quantized hopfield network for online-continual memory. Nat Commun 2024; 15:3722. [PMID: 38697981 PMCID: PMC11065890 DOI: 10.1038/s41467-024-46976-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/21/2023] [Accepted: 03/13/2024] [Indexed: 05/05/2024] Open
Abstract
An important difference between brains and deep neural networks is the way they learn. Nervous systems learn online where a stream of noisy data points are presented in a non-independent, identically distributed way. Further, synaptic plasticity in the brain depends only on information local to synapses. Deep networks, on the other hand, typically use non-local learning algorithms and are trained in an offline, non-noisy, independent, identically distributed setting. Understanding how neural networks learn under the same constraints as the brain is an open problem for neuroscience and neuromorphic computing. A standard approach to this problem has yet to be established. In this paper, we propose that discrete graphical models that learn via an online maximum a posteriori learning algorithm could provide such an approach. We implement this kind of model in a neural network called the Sparse Quantized Hopfield Network. We show our model outperforms state-of-the-art neural networks on associative memory tasks, outperforms these networks in online, continual settings, learns efficiently with noisy inputs, and is better than baselines on an episodic memory task.
Collapse
Affiliation(s)
- Nicholas Alonso
- Department of Cognitive Science, University of California, Irvine, CA, USA.
| | - Jeffrey L Krichmar
- Department of Cognitive Science, University of California, Irvine, CA, USA
- Department Computer Science, University of California, Irvine, CA, USA
| |
Collapse
|
10
|
Terada Y, Toyoizumi T. Chaotic neural dynamics facilitate probabilistic computations through sampling. Proc Natl Acad Sci U S A 2024; 121:e2312992121. [PMID: 38648479 PMCID: PMC11067032 DOI: 10.1073/pnas.2312992121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Accepted: 02/13/2024] [Indexed: 04/25/2024] Open
Abstract
Cortical neurons exhibit highly variable responses over trials and time. Theoretical works posit that this variability arises potentially from chaotic network dynamics of recurrently connected neurons. Here, we demonstrate that chaotic neural dynamics, formed through synaptic learning, allow networks to perform sensory cue integration in a sampling-based implementation. We show that the emergent chaotic dynamics provide neural substrates for generating samples not only of a static variable but also of a dynamical trajectory, where generic recurrent networks acquire these abilities with a biologically plausible learning rule through trial and error. Furthermore, the networks generalize their experience in the stimulus-evoked samples to the inference without partial or all sensory information, which suggests a computational role of spontaneous activity as a representation of the priors as well as a tractable biological computation for marginal distributions. These findings suggest that chaotic neural dynamics may serve for the brain function as a Bayesian generative model.
Collapse
Affiliation(s)
- Yu Terada
- Laboratory for Neural Computation and Adaptation, RIKEN Center for Brain Science, Saitama351-0198, Japan
- Department of Neurobiology, University of California, San Diego, La Jolla, CA92093
- The Institute for Physics of Intelligence, The University of Tokyo, Tokyo113-0033, Japan
| | - Taro Toyoizumi
- Laboratory for Neural Computation and Adaptation, RIKEN Center for Brain Science, Saitama351-0198, Japan
- Department of Mathematical Informatics, Graduate School of Information Science and Technology, The University of Tokyo, Tokyo113-8656, Japan
| |
Collapse
|
11
|
Menéndez JA, Hennig JA, Golub MD, Oby ER, Sadtler PT, Batista AP, Chase SM, Yu BM, Latham PE. A theory of brain-computer interface learning via low-dimensional control. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.04.18.589952. [PMID: 38712193 PMCID: PMC11071278 DOI: 10.1101/2024.04.18.589952] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2024]
Abstract
A remarkable demonstration of the flexibility of mammalian motor systems is primates' ability to learn to control brain-computer interfaces (BCIs). This constitutes a completely novel motor behavior, yet primates are capable of learning to control BCIs under a wide range of conditions. BCIs with carefully calibrated decoders, for example, can be learned with only minutes to hours of practice. With a few weeks of practice, even BCIs with randomly constructed decoders can be learned. What are the biological substrates of this learning process? Here, we develop a theory based on a re-aiming strategy, whereby learning operates within a low-dimensional subspace of task-relevant inputs driving the local population of recorded neurons. Through comprehensive numerical and formal analysis, we demonstrate that this theory can provide a unifying explanation for disparate phenomena previously reported in three different BCI learning tasks, and we derive a novel experimental prediction that we verify with previously published data. By explicitly modeling the underlying neural circuitry, the theory reveals an interpretation of these phenomena in terms of biological constraints on neural activity.
Collapse
|
12
|
Wang Y, Wang Y, Zhang X, Du J, Zhang T, Xu B. Brain topology improved spiking neural network for efficient reinforcement learning of continuous control. Front Neurosci 2024; 18:1325062. [PMID: 38694900 PMCID: PMC11062182 DOI: 10.3389/fnins.2024.1325062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 03/27/2024] [Indexed: 05/04/2024] Open
Abstract
The brain topology highly reflects the complex cognitive functions of the biological brain after million-years of evolution. Learning from these biological topologies is a smarter and easier way to achieve brain-like intelligence with features of efficiency, robustness, and flexibility. Here we proposed a brain topology-improved spiking neural network (BT-SNN) for efficient reinforcement learning. First, hundreds of biological topologies are generated and selected as subsets of the Allen mouse brain topology with the help of the Tanimoto hierarchical clustering algorithm, which has been widely used in analyzing key features of the brain connectome. Second, a few biological constraints are used to filter out three key topology candidates, including but not limited to the proportion of node functions (e.g., sensation, memory, and motor types) and network sparsity. Third, the network topology is integrated with the hybrid numerical solver-improved leaky-integrated and fire neurons. Fourth, the algorithm is then tuned with an evolutionary algorithm named adaptive random search instead of backpropagation to guide synaptic modifications without affecting raw key features of the topology. Fifth, under the test of four animal-survival-like RL tasks (i.e., dynamic controlling in Mujoco), the BT-SNN can achieve higher scores than not only counterpart SNN using random topology but also some classical ANNs (i.e., long-short-term memory and multi-layer perception). This result indicates that the research effort of incorporating biological topology and evolutionary learning rules has much in store for the future.
Collapse
Affiliation(s)
- Yongjian Wang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Yansong Wang
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Future Technology, University of Chinese Academy of Sciences, Beijing, China
| | - Xinhe Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Jiulin Du
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Life Science and Technology, ShanghaiTech University, Shanghai, China
| | - Tielin Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
| | - Bo Xu
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Future Technology, University of Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
13
|
Karimi A, Kalhor A, Sadeghi Tabrizi M. Forward layer-wise learning of convolutional neural networks through separation index maximizing. Sci Rep 2024; 14:8576. [PMID: 38615041 DOI: 10.1038/s41598-024-59176-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2023] [Accepted: 04/08/2024] [Indexed: 04/15/2024] Open
Abstract
This paper proposes a forward layer-wise learning algorithm for CNNs in classification problems. The algorithm utilizes the Separation Index (SI) as a supervised complexity measure to evaluate and train each layer in a forward manner. The proposed method explains that gradually increasing the SI through layers reduces the input data's uncertainties and disturbances, achieving a better feature space representation. Hence, by approximating the SI with a variant of local triplet loss at each layer, a gradient-based learning algorithm is suggested to maximize it. Inspired by the NGRAD (Neural Gradient Representation by Activity Differences) hypothesis, the proposed algorithm operates in a forward manner without explicit error information from the last layer. The algorithm's performance is evaluated on image classification tasks using VGG16, VGG19, AlexNet, and LeNet architectures with CIFAR-10, CIFAR-100, Raabin-WBC, and Fashion-MNIST datasets. Additionally, the experiments are applied to text classification tasks using the DBPedia and AG's News datasets. The results demonstrate that the proposed layer-wise learning algorithm outperforms state-of-the-art methods in accuracy and time complexity.
Collapse
Affiliation(s)
- Ali Karimi
- School of Electrical and Computer Engineering, College of Engineering, University of Tehran, Tehran, Iran
| | - Ahmad Kalhor
- School of Electrical and Computer Engineering, College of Engineering, University of Tehran, Tehran, Iran.
| | - Melika Sadeghi Tabrizi
- School of Electrical and Computer Engineering, College of Engineering, University of Tehran, Tehran, Iran
| |
Collapse
|
14
|
Lee K, Dora S, Mejias JF, Bohte SM, Pennartz CMA. Predictive coding with spiking neurons and feedforward gist signaling. Front Comput Neurosci 2024; 18:1338280. [PMID: 38680678 PMCID: PMC11045951 DOI: 10.3389/fncom.2024.1338280] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Accepted: 03/14/2024] [Indexed: 05/01/2024] Open
Abstract
Predictive coding (PC) is an influential theory in neuroscience, which suggests the existence of a cortical architecture that is constantly generating and updating predictive representations of sensory inputs. Owing to its hierarchical and generative nature, PC has inspired many computational models of perception in the literature. However, the biological plausibility of existing models has not been sufficiently explored due to their use of artificial neurons that approximate neural activity with firing rates in the continuous time domain and propagate signals synchronously. Therefore, we developed a spiking neural network for predictive coding (SNN-PC), in which neurons communicate using event-driven and asynchronous spikes. Adopting the hierarchical structure and Hebbian learning algorithms from previous PC neural network models, SNN-PC introduces two novel features: (1) a fast feedforward sweep from the input to higher areas, which generates a spatially reduced and abstract representation of input (i.e., a neural code for the gist of a scene) and provides a neurobiological alternative to an arbitrary choice of priors; and (2) a separation of positive and negative error-computing neurons, which counters the biological implausibility of a bi-directional error neuron with a very high baseline firing rate. After training with the MNIST handwritten digit dataset, SNN-PC developed hierarchical internal representations and was able to reconstruct samples it had not seen during training. SNN-PC suggests biologically plausible mechanisms by which the brain may perform perceptual inference and learning in an unsupervised manner. In addition, it may be used in neuromorphic applications that can utilize its energy-efficient, event-driven, local learning, and parallel information processing nature.
Collapse
Affiliation(s)
- Kwangjun Lee
- Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, Amsterdam, Netherlands
| | - Shirin Dora
- Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, Amsterdam, Netherlands
- Department of Computer Science, School of Science, Loughborough University, Loughborough, United Kingdom
| | - Jorge F. Mejias
- Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, Amsterdam, Netherlands
| | - Sander M. Bohte
- Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, Amsterdam, Netherlands
- Machine Learning Group, Centre of Mathematics and Computer Science, Amsterdam, Netherlands
| | - Cyriel M. A. Pennartz
- Cognitive and Systems Neuroscience Group, Swammerdam Institute for Life Sciences, Faculty of Science, University of Amsterdam, Amsterdam, Netherlands
| |
Collapse
|
15
|
Juliani A, Safron A, Kanai R. Deep CANALs: a deep learning approach to refining the canalization theory of psychopathology. Neurosci Conscious 2024; 2024:niae005. [PMID: 38533457 PMCID: PMC10965250 DOI: 10.1093/nc/niae005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2023] [Revised: 01/15/2024] [Accepted: 01/19/2024] [Indexed: 03/28/2024] Open
Abstract
Psychedelic therapy has seen a resurgence of interest in the last decade, with promising clinical outcomes for the treatment of a variety of psychopathologies. In response to this success, several theoretical models have been proposed to account for the positive therapeutic effects of psychedelics. One of the more prominent models is "RElaxed Beliefs Under pSychedelics," which proposes that psychedelics act therapeutically by relaxing the strength of maladaptive high-level beliefs encoded in the brain. The more recent "CANAL" model of psychopathology builds on the explanatory framework of RElaxed Beliefs Under pSychedelics by proposing that canalization (the development of overly rigid belief landscapes) may be a primary factor in psychopathology. Here, we make use of learning theory in deep neural networks to develop a series of refinements to the original CANAL model. Our primary theoretical contribution is to disambiguate two separate optimization landscapes underlying belief representation in the brain and describe the unique pathologies which can arise from the canalization of each. Along each dimension, we identify pathologies of either too much or too little canalization, implying that the construct of canalization does not have a simple linear correlation with the presentation of psychopathology. In this expanded paradigm, we demonstrate the ability to make novel predictions regarding what aspects of psychopathology may be amenable to psychedelic therapy, as well as what forms of psychedelic therapy may ultimately be most beneficial for a given individual.
Collapse
Affiliation(s)
- Arthur Juliani
- Microsoft Research , Microsoft, 300 Lafayette St, New York, NY 10012, USA
| | - Adam Safron
- Department of Psychiatry and Behavioral Sciences, Johns Hopkins University, 600 N Wolfe St, Baltimore, MD 21205, USA
| | - Ryota Kanai
- Neurotechnology R & D Unit, Araya Inc, 6F Sanpo Sakuma Building, 1-11 Kandasakumacho, Chiyoda-ku, Tokyo 101-0025, Japan
| |
Collapse
|
16
|
Wolff M, Halassa MM. The mediodorsal thalamus in executive control. Neuron 2024; 112:893-908. [PMID: 38295791 DOI: 10.1016/j.neuron.2024.01.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 11/15/2023] [Accepted: 01/03/2024] [Indexed: 03/23/2024]
Abstract
Executive control, the ability to organize thoughts and action plans in real time, is a defining feature of higher cognition. Classical theories have emphasized cortical contributions to this process, but recent studies have reinvigorated interest in the role of the thalamus. Although it is well established that local thalamic damage diminishes cognitive capacity, such observations have been difficult to inform functional models. Recent progress in experimental techniques is beginning to enrich our understanding of the anatomical, physiological, and computational substrates underlying thalamic engagement in executive control. In this review, we discuss this progress and particularly focus on the mediodorsal thalamus, which regulates the activity within and across frontal cortical areas. We end with a synthesis that highlights frontal thalamocortical interactions in cognitive computations and discusses its functional implications in normal and pathological conditions.
Collapse
Affiliation(s)
- Mathieu Wolff
- University of Bordeaux, CNRS, INCIA, UMR 5287, 33000 Bordeaux, France.
| | - Michael M Halassa
- Department of Neuroscience, Tufts University School of Medicine, Boston, MA, USA; Department of Psychiatry, Tufts University School of Medicine, Boston, MA, USA.
| |
Collapse
|
17
|
Bi Z, Li H, Tian L. Top-down generation of low-resolution representations improves visual perception and imagination. Neural Netw 2024; 171:440-456. [PMID: 38150870 DOI: 10.1016/j.neunet.2023.12.030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2023] [Revised: 11/30/2023] [Accepted: 12/18/2023] [Indexed: 12/29/2023]
Abstract
Perception or imagination requires top-down signals from high-level cortex to primary visual cortex (V1) to reconstruct or simulate the representations bottom-up stimulated by the seen images. Interestingly, top-down signals in V1 have lower spatial resolution than bottom-up representations. It is unclear why the brain uses low-resolution signals to reconstruct or simulate high-resolution representations. By modeling the top-down pathway of the visual system using the decoder of a variational auto-encoder (VAE), we reveal that low-resolution top-down signals can better reconstruct or simulate the information contained in the sparse activities of V1 simple cells, which facilitates perception and imagination. This advantage of low-resolution generation is related to facilitating high-level cortex to form geometry-respecting representations observed in experiments. Furthermore, we present two findings regarding this phenomenon in the context of AI-generated sketches, a style of drawings made of lines. First, we found that the quality of the generated sketches critically depends on the thickness of the lines in the sketches: thin-line sketches are harder to generate than thick-line sketches. Second, we propose a technique to generate high-quality thin-line sketches: instead of directly using original thin-line sketches, we use blurred sketches to train VAE or GAN (generative adversarial network), and then infer the thin-line sketches from the VAE- or GAN-generated blurred sketches. Collectively, our work suggests that low-resolution top-down generation is a strategy the brain uses to improve visual perception and imagination, which inspires new sketch-generation AI techniques.
Collapse
Affiliation(s)
- Zedong Bi
- Lingang Laboratory, Shanghai 200031, China.
| | - Haoran Li
- Department of Physics, Hong Kong Baptist University, Hong Kong, China
| | - Liang Tian
- Department of Physics, Hong Kong Baptist University, Hong Kong, China; Institute of Computational and Theoretical Studies, Hong Kong Baptist University, Hong Kong, China; Institute of Systems Medicine and Health Sciences, Hong Kong Baptist University, Hong Kong, China; State Key Laboratory of Environmental and Biological Analysis, Hong Kong Baptist University, Hong Kong, China.
| |
Collapse
|
18
|
Chang CK, Yang CW. Retrieving profile of photoresist with high aspect ratio and subwavelength features using optical spectroscopy and artificial neural network. OPTICS EXPRESS 2024; 32:8389-8396. [PMID: 38439495 DOI: 10.1364/oe.517201] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 02/13/2024] [Indexed: 03/06/2024]
Abstract
Profile measurements of structures with a high aspect ratio and subwavelength features (HARSW) can be achieved using transmission electron microscopy and tilted scanning electron microscopy. Although electron microscopy can provide accurate HARSW measurements, it is laborious and destructive. In this paper, nondestructive and labor-saving methods were proposed to measure the dimensions of HARSW structures. The optical reflection spectrum, along with an artificial neural network (ANN) model, was adopted for interpolation with the simulation database to retrieve the dimensions of HARSW structures. To generate the ANN model, the experimental and simulated reflection spectra were adopted as the input and output variables for the training data, respectively. This ANN model can learn the discrepancy between simulation and experimental reflections. The finite-difference time-domain method was also adopted to calculate the simulated reflection spectra of HARSW structures with various dimensions, which can be used as a database. Once the experimental reflection of a HARSW structure with unknown dimensions was obtained, the ANN model could generate a simulation-like reflection spectrum. Linear regression was used to determine the correlation coefficients of the simulation-like reflection spectra in the database. The accurate dimensions of HARSW structures can be determined using a higher correlation coefficient. This methodology can be a prominent method for the process monitoring of HARSW structures.
Collapse
|
19
|
Irastorza-Valera L, Benítez JM, Montáns FJ, Saucedo-Mora L. An Agent-Based Model to Reproduce the Boolean Logic Behaviour of Neuronal Self-Organised Communities through Pulse Delay Modulation and Generation of Logic Gates. Biomimetics (Basel) 2024; 9:101. [PMID: 38392147 PMCID: PMC10886514 DOI: 10.3390/biomimetics9020101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2023] [Revised: 01/16/2024] [Accepted: 02/04/2024] [Indexed: 02/24/2024] Open
Abstract
The human brain is arguably the most complex "machine" to ever exist. Its detailed functioning is yet to be fully understood, let alone modelled. Neurological processes have logical signal-processing and biophysical aspects, and both affect the brain's structure, functioning and adaptation. Mathematical approaches based on both information and graph theory have been extensively used in an attempt to approximate its biological functioning, along with Artificial Intelligence frameworks inspired by its logical functioning. In this article, an approach to model some aspects of the brain learning and signal processing is presented, mimicking the metastability and backpropagation found in the real brain while also accounting for neuroplasticity. Several simulations are carried out with this model to demonstrate how dynamic neuroplasticity, neural inhibition and neuron migration can reshape the brain's logical connectivity to synchronise signal processing and obtain certain target latencies. This work showcases the importance of dynamic logical and biophysical remodelling in brain plasticity. Combining mathematical (agents, graph theory, topology and backpropagation) and biomedical ingredients (metastability, neuroplasticity and migration), these preliminary results prove complex brain phenomena can be reproduced-under pertinent simplifications-via affordable computations, which can be construed as a starting point for more ambitiously accurate simulations.
Collapse
Affiliation(s)
- Luis Irastorza-Valera
- E.T.S. de Ingeniería Aeronáutica y del Espacio, Universidad Politécnica de Madrid, Pza. Cardenal Cisneros 3, 28040 Madrid, Spain
- PIMM Laboratory, Arts et Métiers Institute of Technology, 151 Bd de l'Hôpital, 75013 Paris, France
| | - José María Benítez
- E.T.S. de Ingeniería Aeronáutica y del Espacio, Universidad Politécnica de Madrid, Pza. Cardenal Cisneros 3, 28040 Madrid, Spain
| | - Francisco J Montáns
- E.T.S. de Ingeniería Aeronáutica y del Espacio, Universidad Politécnica de Madrid, Pza. Cardenal Cisneros 3, 28040 Madrid, Spain
- Department of Mechanical and Aerospace Engineering, Herbert Wertheim College of Engineering, University of Florida, Gainesville, FL 32611, USA
| | - Luis Saucedo-Mora
- E.T.S. de Ingeniería Aeronáutica y del Espacio, Universidad Politécnica de Madrid, Pza. Cardenal Cisneros 3, 28040 Madrid, Spain
- Department of Materials, University of Oxford, Parks Road, Oxford OX1 3PJ, UK
- Department of Nuclear Science and Engineering, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| |
Collapse
|
20
|
Matteucci G, Piasini E, Zoccolan D. Unsupervised learning of mid-level visual representations. Curr Opin Neurobiol 2024; 84:102834. [PMID: 38154417 DOI: 10.1016/j.conb.2023.102834] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2023] [Revised: 12/03/2023] [Accepted: 12/05/2023] [Indexed: 12/30/2023]
Abstract
Recently, a confluence between trends in neuroscience and machine learning has brought a renewed focus on unsupervised learning, where sensory processing systems learn to exploit the statistical structure of their inputs in the absence of explicit training targets or rewards. Sophisticated experimental approaches have enabled the investigation of the influence of sensory experience on neural self-organization and its synaptic bases. Meanwhile, novel algorithms for unsupervised and self-supervised learning have become increasingly popular both as inspiration for theories of the brain, particularly for the function of intermediate visual cortical areas, and as building blocks of real-world learning machines. Here we review some of these recent developments, placing them in historical context and highlighting some research lines that promise exciting breakthroughs in the near future.
Collapse
Affiliation(s)
- Giulio Matteucci
- Department of Basic Neurosciences, University of Geneva, Geneva, 1206, Switzerland. https://twitter.com/giulio_matt
| | - Eugenio Piasini
- International School for Advanced Studies (SISSA), Trieste, 34136, Italy
| | - Davide Zoccolan
- International School for Advanced Studies (SISSA), Trieste, 34136, Italy.
| |
Collapse
|
21
|
Barradas VR, Koike Y, Schweighofer N. Theoretical limits on the speed of learning inverse models explain the rate of adaptation in arm reaching tasks. Neural Netw 2024; 170:376-389. [PMID: 38029719 DOI: 10.1016/j.neunet.2023.10.049] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Revised: 09/08/2023] [Accepted: 10/30/2023] [Indexed: 12/01/2023]
Abstract
An essential aspect of human motor learning is the formation of inverse models, which map desired actions to motor commands. Inverse models can be learned by adjusting parameters in neural circuits to minimize errors in the performance of motor tasks through gradient descent. However, the theory of gradient descent establishes limits on the learning speed. Specifically, the eigenvalues of the Hessian of the error surface around a minimum determine the maximum speed of learning in a task. Here, we use this theoretical framework to analyze the speed of learning in different inverse model learning architectures in a set of isometric arm-reaching tasks. We show theoretically that, in these tasks, the error surface and, thus the speed of learning, are determined by the shapes of the force manipulability ellipsoid of the arm and the distribution of targets in the task. In particular, rounder manipulability ellipsoids generate a rounder error surface, allowing for faster learning of the inverse model. Rounder target distributions have a similar effect. We tested these predictions experimentally in a quasi-isometric reaching task with a visuomotor transformation. The experimental results were consistent with our theoretical predictions. Furthermore, our analysis accounts for the speed of learning in previous experiments with incompatible and compatible virtual surgery tasks, and with visuomotor rotation tasks with different numbers of targets. By identifying aspects of a task that influence the speed of learning, our results provide theoretical principles for the design of motor tasks that allow for faster learning.
Collapse
Affiliation(s)
- Victor R Barradas
- Institute of Innovative Research, Tokyo Institute of Technology, 4259 R2-16 Nagatsuta-cho, Midori-ku, Yokohama, Kanagawa 226-8503, Japan.
| | - Yasuharu Koike
- Institute of Innovative Research, Tokyo Institute of Technology, 4259 R2-16 Nagatsuta-cho, Midori-ku, Yokohama, Kanagawa 226-8503, Japan
| | - Nicolas Schweighofer
- Biokinesiology and Physical Therapy, University of Southern California, 1540 Alcazar Street, CHP 155, Los Angeles, CA 90089-9006, USA
| |
Collapse
|
22
|
de Brito CSN, Gerstner W. Learning what matters: Synaptic plasticity with invariance to second-order input correlations. PLoS Comput Biol 2024; 20:e1011844. [PMID: 38346073 PMCID: PMC10890752 DOI: 10.1371/journal.pcbi.1011844] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Revised: 02/23/2024] [Accepted: 01/18/2024] [Indexed: 02/25/2024] Open
Abstract
Cortical populations of neurons develop sparse representations adapted to the statistics of the environment. To learn efficient population codes, synaptic plasticity mechanisms must differentiate relevant latent features from spurious input correlations, which are omnipresent in cortical networks. Here, we develop a theory for sparse coding and synaptic plasticity that is invariant to second-order correlations in the input. Going beyond classical Hebbian learning, our learning objective explains the functional form of observed excitatory plasticity mechanisms, showing how Hebbian long-term depression (LTD) cancels the sensitivity to second-order correlations so that receptive fields become aligned with features hidden in higher-order statistics. Invariance to second-order correlations enhances the versatility of biologically realistic learning models, supporting optimal decoding from noisy inputs and sparse population coding from spatially correlated stimuli. In a spiking model with triplet spike-timing-dependent plasticity (STDP), we show that individual neurons can learn localized oriented receptive fields, circumventing the need for input preprocessing, such as whitening, or population-level lateral inhibition. The theory advances our understanding of local unsupervised learning in cortical circuits, offers new interpretations of the Bienenstock-Cooper-Munro and triplet STDP models, and assigns a specific functional role to synaptic LTD mechanisms in pyramidal neurons.
Collapse
Affiliation(s)
- Carlos Stein Naves de Brito
- École Polytechnique Fédérale de Lausanne, EPFL, Lusanne, Switzerland
- Champalimaud Research, Champalimaud Centre for the Unknown, Lisbon, Portugal
| | - Wulfram Gerstner
- École Polytechnique Fédérale de Lausanne, EPFL, Lusanne, Switzerland
| |
Collapse
|
23
|
Deperrois N, Petrovici MA, Senn W, Jordan J. Learning beyond sensations: How dreams organize neuronal representations. Neurosci Biobehav Rev 2024; 157:105508. [PMID: 38097096 DOI: 10.1016/j.neubiorev.2023.105508] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Revised: 12/05/2023] [Accepted: 12/09/2023] [Indexed: 12/25/2023]
Abstract
Semantic representations in higher sensory cortices form the basis for robust, yet flexible behavior. These representations are acquired over the course of development in an unsupervised fashion and continuously maintained over an organism's lifespan. Predictive processing theories propose that these representations emerge from predicting or reconstructing sensory inputs. However, brains are known to generate virtual experiences, such as during imagination and dreaming, that go beyond previously experienced inputs. Here, we suggest that virtual experiences may be just as relevant as actual sensory inputs in shaping cortical representations. In particular, we discuss two complementary learning principles that organize representations through the generation of virtual experiences. First, "adversarial dreaming" proposes that creative dreams support a cortical implementation of adversarial learning in which feedback and feedforward pathways engage in a productive game of trying to fool each other. Second, "contrastive dreaming" proposes that the invariance of neuronal representations to irrelevant factors of variation is acquired by trying to map similar virtual experiences together via a contrastive learning process. These principles are compatible with known cortical structure and dynamics and the phenomenology of sleep thus providing promising directions to explain cortical learning beyond the classical predictive processing paradigm.
Collapse
Affiliation(s)
| | | | - Walter Senn
- Department of Physiology, University of Bern, Bern, Switzerland
| | - Jakob Jordan
- Department of Physiology, University of Bern, Bern, Switzerland; Electrical Engineering, Yale University, New Haven, CT, United States
| |
Collapse
|
24
|
Song Y, Millidge B, Salvatori T, Lukasiewicz T, Xu Z, Bogacz R. Inferring neural activity before plasticity as a foundation for learning beyond backpropagation. Nat Neurosci 2024; 27:348-358. [PMID: 38172438 PMCID: PMC7615830 DOI: 10.1038/s41593-023-01514-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Accepted: 11/02/2023] [Indexed: 01/05/2024]
Abstract
For both humans and machines, the essence of learning is to pinpoint which components in its information processing pipeline are responsible for an error in its output, a challenge that is known as 'credit assignment'. It has long been assumed that credit assignment is best solved by backpropagation, which is also the foundation of modern machine learning. Here, we set out a fundamentally different principle on credit assignment called 'prospective configuration'. In prospective configuration, the network first infers the pattern of neural activity that should result from learning, and then the synaptic weights are modified to consolidate the change in neural activity. We demonstrate that this distinct mechanism, in contrast to backpropagation, (1) underlies learning in a well-established family of models of cortical circuits, (2) enables learning that is more efficient and effective in many contexts faced by biological organisms and (3) reproduces surprising patterns of neural activity and behavior observed in diverse human and rat learning experiments.
Collapse
Affiliation(s)
- Yuhang Song
- Department of Computer Science, University of Oxford, Oxford, UK.
- Medical Research Council Brain Network Dynamics Unit, University of Oxford, Oxford, UK.
- Fractile, Ltd., London, UK.
| | - Beren Millidge
- Medical Research Council Brain Network Dynamics Unit, University of Oxford, Oxford, UK
| | - Tommaso Salvatori
- Department of Computer Science, University of Oxford, Oxford, UK
- Institute of Logic and Computation, Vienna University of Technology, Vienna, Austria
- VERSES AI Research Lab, Los Angeles, CA, USA
| | - Thomas Lukasiewicz
- Department of Computer Science, University of Oxford, Oxford, UK.
- Institute of Logic and Computation, Vienna University of Technology, Vienna, Austria.
| | - Zhenghua Xu
- Department of Computer Science, University of Oxford, Oxford, UK.
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, School of Health Sciences and Biomedical Engineering, Hebei University of Technology, Tianjin, China.
| | - Rafal Bogacz
- Medical Research Council Brain Network Dynamics Unit, University of Oxford, Oxford, UK.
| |
Collapse
|
25
|
Joshi S, Haney S, Wang Z, Locatelli F, Smith B, Cao Y, Bazhenov M. Plasticity in inhibitory networks improves pattern separation in early olfactory processing. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.01.24.576675. [PMID: 38328149 PMCID: PMC10849730 DOI: 10.1101/2024.01.24.576675] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/09/2024]
Abstract
Distinguishing between nectar and non-nectar odors presents a challenge for animals due to shared compounds in complex mixtures, where changing ratios often signify differences in reward. Changes in nectar production throughout the day and potentially many times within a forager's lifetime add to the complexity. The honeybee olfactory system, containing less than a 1000 of principal neurons in the early olfactory relay, the antennal lobe (AL), must learn to associate diverse volatile blends with rewards. We used a computational network model and live imaging of the honeybee's AL to explore the neural mechanisms and functions of the AL plasticity. Our findings revealed that when trained with a set of rewarded and unrewarded odors, the AL inhibitory network suppresses shared chemical compounds while enhancing responses to distinct compounds. This results in improved pattern separation and a more concise and efficient neural code. Our Ca2+ imaging data support our model's predictions. Furthermore, we applied these contrast enhancement principles to a Graph Convolutional Network (GCN) and found that similar mechanisms could enhance the performance of artificial neural networks. Our model provides insights into how plasticity at the inhibitory network level reshapes coding for efficient learning of complex odors.
Collapse
Affiliation(s)
- Shruti Joshi
- Department of Electrical and Computer Engineering, University of California San Diego, USA
- Department of Medicine, University of California San Diego, USA
| | - Seth Haney
- Department of Medicine, University of California San Diego, USA
| | - Zhenyu Wang
- Department of Electrical, Computer and Energy Engineering, Arizona State University, USA
| | - Fernando Locatelli
- Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires, Instituto de Fisiología, Biología Molecular y Neurociencias, CONICET, Buenos Aires, Argentina
| | - Brian Smith
- School of Life Science, Arizona State University, USA
| | - Yu Cao
- Department of Electrical and Computer Engineering, University of Minnesota, USA
| | - Maxim Bazhenov
- Department of Medicine, University of California San Diego, USA
| |
Collapse
|
26
|
Bonanno GA, Chen S, Bagrodia R, Galatzer-Levy IR. Resilience and Disaster: Flexible Adaptation in the Face of Uncertain Threat. Annu Rev Psychol 2024; 75:573-599. [PMID: 37566760 DOI: 10.1146/annurev-psych-011123-024224] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/13/2023]
Abstract
Disasters cause sweeping damage, hardship, and loss of life. In this article, we first consider the dominant psychological approach to disasters and its narrow focus on psychopathology (e.g., posttraumatic stress disorder). We then review research on a broader approach that has identified heterogeneous, highly replicable trajectories of outcome, the most common being stable mental health or resilience. We review trajectory research for different types of disasters, including the COVID-19 pandemic. Next, we consider correlates of the resilience trajectory and note their paradoxically limited ability to predict future resilient outcomes. Research using machine learning algorithms improved prediction but has not yet illuminated the mechanism behind resilient adaptation. To that end, we propose a more direct psychological explanation for resilience based on research on the motivational and mechanistic components of regulatory flexibility. Finally, we consider how future research might leverage new computational approaches to better capture regulatory flexibility in real time.
Collapse
Affiliation(s)
- George A Bonanno
- Department of Counseling and Clinical Psychology, Teachers College, Columbia University, New York, NY, USA; , ,
| | - Shuquan Chen
- Department of Counseling and Clinical Psychology, Teachers College, Columbia University, New York, NY, USA; , ,
| | - Rohini Bagrodia
- Department of Counseling and Clinical Psychology, Teachers College, Columbia University, New York, NY, USA; , ,
| | - Isaac R Galatzer-Levy
- Department of Psychiatry, Grossman School of Medicine, New York University, New York, NY, USA;
- Google LLC, Mountain View, California
| |
Collapse
|
27
|
Fernandez-Leon JA, Sarramone L. The grid-cell normative model: Unifying 'principles'. Biosystems 2024; 235:105091. [PMID: 38040283 DOI: 10.1016/j.biosystems.2023.105091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Revised: 11/21/2023] [Accepted: 11/21/2023] [Indexed: 12/03/2023]
Abstract
A normative model for the emergence of entorhinal grid cells in the brain's navigational system has been proposed (Sorscher et al., 2023. Neuron 111, 121-137). Using computational modeling of place-to-grid cell interactions, the authors characterized the fundamental nature of grid cells through information processing. However, the normative model does not consider certain discoveries that complement or contradict the conditions for such emergence. By briefly reviewing current evidence, we draw some implications on the interplay between place cell replay sequences and intrinsic grid cell oscillations related to the hippocampal-entorhinal navigation system that can extend the normative model.
Collapse
Affiliation(s)
- Jose A Fernandez-Leon
- Universidad Nacional del Centro de la Provincia de Buenos Aires (UNCPBA), Fac. Cs. Exactas, INTIA, Tandil, Argentina; Consejo Nacional de Investigaciones Científicas y Técnicas (CONICET), Buenos Aires, Argentina; CIFICEN, UNCPBA-CICPBA-CONICET, Tandil, Argentina.
| | - Luca Sarramone
- Universidad Nacional del Centro de la Provincia de Buenos Aires (UNCPBA), Fac. Cs. Exactas, INTIA, Tandil, Argentina; Consejo Nacional de Investigaciones Científicas y Técnicas (CONICET), Buenos Aires, Argentina
| |
Collapse
|
28
|
Bi Z. Cognition of Time and Thinking Beyond. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2024; 1455:171-195. [PMID: 38918352 DOI: 10.1007/978-3-031-60183-5_10] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/27/2024]
Abstract
A common research protocol in cognitive neuroscience is to train subjects to perform deliberately designed experiments while recording brain activity, with the aim of understanding the brain mechanisms underlying cognition. However, how the results of this protocol of research can be applied in technology is seldom discussed. Here, I review the studies on time processing of the brain as examples of this research protocol, as well as two main application areas of neuroscience (neuroengineering and brain-inspired artificial intelligence). Time processing is a fundamental dimension of cognition, and time is also an indispensable dimension of any real-world signal to be processed in technology. Therefore, one may expect that the studies of time processing in cognition profoundly influence brain-related technology. Surprisingly, I found that the results from cognitive studies on timing processing are hardly helpful in solving practical problems. This awkward situation may be due to the lack of generalizability of the results of cognitive studies, which are under well-controlled laboratory conditions, to real-life situations. This lack of generalizability may be rooted in the fundamental unknowability of the world (including cognition). Overall, this paper questions and criticizes the usefulness and prospect of the abovementioned research protocol of cognitive neuroscience. I then give three suggestions for future research. First, to improve the generalizability of research, it is better to study brain activity under real-life conditions instead of in well-controlled laboratory experiments. Second, to overcome the unknowability of the world, we can engineer an easily accessible surrogate of the object under investigation, so that we can predict the behavior of the object under investigation by experimenting on the surrogate. Third, the paper calls for technology-oriented research, with the aim of technology creation instead of knowledge discovery.
Collapse
Affiliation(s)
- Zedong Bi
- Lingang Laboratory, Shanghai, China.
- Institute for Future, Qingdao University, Qingdao, China.
- School of Automation, Shandong Key Laboratory of Industrial Control Technology, Qingdao University, Qingdao, China.
| |
Collapse
|
29
|
Liu Z, Gan E, Tegmark M. Seeing Is Believing: Brain-Inspired Modular Training for Mechanistic Interpretability. ENTROPY (BASEL, SWITZERLAND) 2023; 26:41. [PMID: 38248167 PMCID: PMC10814460 DOI: 10.3390/e26010041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Revised: 12/21/2023] [Accepted: 12/27/2023] [Indexed: 01/23/2024]
Abstract
We introduce Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable. Inspired by brains, BIMT embeds neurons in a geometric space and augments the loss function with a cost proportional to the length of each neuron connection. This is inspired by the idea of minimum connection cost in evolutionary biology, but we are the first the combine this idea with training neural networks with gradient descent for interpretability. We demonstrate that BIMT discovers useful modular neural networks for many simple tasks, revealing compositional structures in symbolic formulas, interpretable decision boundaries and features for classification, and mathematical structure in algorithmic datasets. Qualitatively, BIMT-trained networks have modules readily identifiable by the naked eye, but regularly trained networks seem much more complicated. Quantitatively, we use Newman's method to compute the modularity of network graphs; BIMT achieves the highest modularity for all our test problems. A promising and ambitious future direction is to apply the proposed method to understand large models for vision, language, and science.
Collapse
Affiliation(s)
- Ziming Liu
- Institute for Artificial Intelligence and Fundamental Interactions, Massachusetts Institute of Technology, Cambridge, MA 02139, USA; (E.G.); (M.T.)
| | | | | |
Collapse
|
30
|
Rao RPN, Gklezakos DC, Sathish V. Active Predictive Coding: A Unifying Neural Model for Active Perception, Compositional Learning, and Hierarchical Planning. Neural Comput 2023; 36:1-32. [PMID: 38052084 DOI: 10.1162/neco_a_01627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 09/20/2023] [Indexed: 12/07/2023]
Abstract
There is growing interest in predictive coding as a model of how the brain learns through predictions and prediction errors. Predictive coding models have traditionally focused on sensory coding and perception. Here we introduce active predictive coding (APC) as a unifying model for perception, action, and cognition. The APC model addresses important open problems in cognitive science and AI, including (1) how we learn compositional representations (e.g., part-whole hierarchies for equivariant vision) and (2) how we solve large-scale planning problems, which are hard for traditional reinforcement learning, by composing complex state dynamics and abstract actions from simpler dynamics and primitive actions. By using hypernetworks, self-supervised learning, and reinforcement learning, APC learns hierarchical world models by combining task-invariant state transition networks and task-dependent policy networks at multiple abstraction levels. We illustrate the applicability of the APC model to active visual perception and hierarchical planning. Our results represent, to our knowledge, the first proof-of-concept demonstration of a unified approach to addressing the part-whole learning problem in vision, the nested reference frames learning problem in cognition, and the integrated state-action hierarchy learning problem in reinforcement learning.
Collapse
Affiliation(s)
- Rajesh P N Rao
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| | - Dimitrios C Gklezakos
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| | - Vishwas Sathish
- Paul G. Allen School of Computer Science and Engineering and Center for Neurotechnology, University of Washington, Seattle, WA 98195, U.S.A.
| |
Collapse
|
31
|
Linsley D, Serre T. Fixing the problems of deep neural networks will require better training data and learning algorithms. Behav Brain Sci 2023; 46:e400. [PMID: 38054333 DOI: 10.1017/s0140525x23001589] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
Bowers et al. argue that deep neural networks (DNNs) are poor models of biological vision because they often learn to rival human accuracy by relying on strategies that differ markedly from those of humans. We show that this problem is worsening as DNNs are becoming larger-scale and increasingly more accurate, and prescribe methods for building DNNs that can reliably model biological vision.
Collapse
Affiliation(s)
- Drew Linsley
- Department of Cognitive Linguistic & Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, RI, USA ://sites.brown.edu/drewlinsleyhttps://serre-lab.clps.brown.edu
| | - Thomas Serre
- Department of Cognitive Linguistic & Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, RI, USA ://sites.brown.edu/drewlinsleyhttps://serre-lab.clps.brown.edu
| |
Collapse
|
32
|
Suzuki M, Pennartz CMA, Aru J. How deep is the brain? The shallow brain hypothesis. Nat Rev Neurosci 2023; 24:778-791. [PMID: 37891398 DOI: 10.1038/s41583-023-00756-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/25/2023] [Indexed: 10/29/2023]
Abstract
Deep learning and predictive coding architectures commonly assume that inference in neural networks is hierarchical. However, largely neglected in deep learning and predictive coding architectures is the neurobiological evidence that all hierarchical cortical areas, higher or lower, project to and receive signals directly from subcortical areas. Given these neuroanatomical facts, today's dominance of cortico-centric, hierarchical architectures in deep learning and predictive coding networks is highly questionable; such architectures are likely to be missing essential computational principles the brain uses. In this Perspective, we present the shallow brain hypothesis: hierarchical cortical processing is integrated with a massively parallel process to which subcortical areas substantially contribute. This shallow architecture exploits the computational capacity of cortical microcircuits and thalamo-cortical loops that are not included in typical hierarchical deep learning and predictive coding networks. We argue that the shallow brain architecture provides several critical benefits over deep hierarchical structures and a more complete depiction of how mammalian brains achieve fast and flexible computational capabilities.
Collapse
Affiliation(s)
- Mototaka Suzuki
- Department of Cognitive and Systems Neuroscience, Swammerdam Institute for Life Sciences, University of Amsterdam, Amsterdam, The Netherlands.
| | - Cyriel M A Pennartz
- Department of Cognitive and Systems Neuroscience, Swammerdam Institute for Life Sciences, University of Amsterdam, Amsterdam, The Netherlands
| | - Jaan Aru
- Institute of Computer Science, University of Tartu, Tartu, Estonia.
| |
Collapse
|
33
|
Friedenberger Z, Harkin E, Tóth K, Naud R. Silences, spikes and bursts: Three-part knot of the neural code. J Physiol 2023; 601:5165-5193. [PMID: 37889516 DOI: 10.1113/jp281510] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2023] [Accepted: 09/28/2023] [Indexed: 10/28/2023] Open
Abstract
When a neuron breaks silence, it can emit action potentials in a number of patterns. Some responses are so sudden and intense that electrophysiologists felt the need to single them out, labelling action potentials emitted at a particularly high frequency with a metonym - bursts. Is there more to bursts than a figure of speech? After all, sudden bouts of high-frequency firing are expected to occur whenever inputs surge. The burst coding hypothesis advances that the neural code has three syllables: silences, spikes and bursts. We review evidence supporting this ternary code in terms of devoted mechanisms for burst generation, synaptic transmission and synaptic plasticity. We also review the learning and attention theories for which such a triad is beneficial.
Collapse
Affiliation(s)
- Zachary Friedenberger
- Brain and Mind Institute, Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, Ontario, Canada
- Centre for Neural Dynamics and Artifical Intelligence, Department of Physics, University of Ottawa, Ottawa, Ontario, Ottawa
| | - Emerson Harkin
- Brain and Mind Institute, Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, Ontario, Canada
| | - Katalin Tóth
- Brain and Mind Institute, Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, Ontario, Canada
| | - Richard Naud
- Brain and Mind Institute, Department of Cellular and Molecular Medicine, University of Ottawa, Ottawa, Ontario, Canada
- Centre for Neural Dynamics and Artifical Intelligence, Department of Physics, University of Ottawa, Ottawa, Ontario, Ottawa
| |
Collapse
|
34
|
Lee MJ, DiCarlo JJ. How well do rudimentary plasticity rules predict adult visual object learning? PLoS Comput Biol 2023; 19:e1011713. [PMID: 38079444 PMCID: PMC10754461 DOI: 10.1371/journal.pcbi.1011713] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Revised: 12/28/2023] [Accepted: 11/27/2023] [Indexed: 12/29/2023] Open
Abstract
A core problem in visual object learning is using a finite number of images of a new object to accurately identify that object in future, novel images. One longstanding, conceptual hypothesis asserts that this core problem is solved by adult brains through two connected mechanisms: 1) the re-representation of incoming retinal images as points in a fixed, multidimensional neural space, and 2) the optimization of linear decision boundaries in that space, via simple plasticity rules applied to a single downstream layer. Though this scheme is biologically plausible, the extent to which it explains learning behavior in humans has been unclear-in part because of a historical lack of image-computable models of the putative neural space, and in part because of a lack of measurements of human learning behaviors in difficult, naturalistic settings. Here, we addressed these gaps by 1) drawing from contemporary, image-computable models of the primate ventral visual stream to create a large set of testable learning models (n = 2,408 models), and 2) using online psychophysics to measure human learning trajectories over a varied set of tasks involving novel 3D objects (n = 371,000 trials), which we then used to develop (and publicly release) empirical benchmarks for comparing learning models to humans. We evaluated each learning model on these benchmarks, and found those based on deep, high-level representations from neural networks were surprisingly aligned with human behavior. While no tested model explained the entirety of replicable human behavior, these results establish that rudimentary plasticity rules, when combined with appropriate visual representations, have high explanatory power in predicting human behavior with respect to this core object learning problem.
Collapse
Affiliation(s)
- Michael J. Lee
- Department of Brain and Cognitive Sciences, MIT, Cambridge, Massachusetts, United States of America
- Center for Brains, Minds and Machines, MIT, Cambridge, Massachusetts, United States of America
| | - James J. DiCarlo
- Department of Brain and Cognitive Sciences, MIT, Cambridge, Massachusetts, United States of America
- Center for Brains, Minds and Machines, MIT, Cambridge, Massachusetts, United States of America
- McGovern Institute for Brain Research, MIT, Cambridge, Massachusetts, United States of America
| |
Collapse
|
35
|
Momeni A, Rahmani B, Malléjac M, Del Hougne P, Fleury R. Backpropagation-free training of deep physical neural networks. Science 2023:eadi8474. [PMID: 37995209 DOI: 10.1126/science.adi8474] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 11/07/2023] [Indexed: 11/25/2023]
Abstract
Recent successes in deep learning for vision and natural language processing are attributed to larger models but come with energy consumption and scalability issues. Current training of digital deep learning models primarily relies on backpropagation that is unsuitable for physical implementation. Here, we proposed a simple deep neural network architecture augmented by a physical local learning (PhyLL) algorithm, enabling supervised and unsupervised training of deep physical neural networks, without detailed knowledge of the nonlinear physical layer's properties. We trained diverse wave-based physical neural networks in vowel and image classification experiments, showcasing our approach's universality. Our method shows advantages over other hardware-aware training schemes by improving training speed, enhancing robustness, and reducing power consumption by eliminating the need for system modelling and thus decreasing digital computation.
Collapse
Affiliation(s)
- Ali Momeni
- Laboratory of Wave Engineering, Department of Electrical Engineering, EPFL, Lausanne CH-1015, Switzerland
| | | | - Matthieu Malléjac
- Laboratory of Wave Engineering, Department of Electrical Engineering, EPFL, Lausanne CH-1015, Switzerland
| | | | - Romain Fleury
- Laboratory of Wave Engineering, Department of Electrical Engineering, EPFL, Lausanne CH-1015, Switzerland
| |
Collapse
|
36
|
Fitch WT. Cellular computation and cognition. Front Comput Neurosci 2023; 17:1107876. [PMID: 38077750 PMCID: PMC10702520 DOI: 10.3389/fncom.2023.1107876] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Accepted: 10/09/2023] [Indexed: 05/28/2024] Open
Abstract
Contemporary neural network models often overlook a central biological fact about neural processing: that single neurons are themselves complex, semi-autonomous computing systems. Both the information processing and information storage abilities of actual biological neurons vastly exceed the simple weighted sum of synaptic inputs computed by the "units" in standard neural network models. Neurons are eukaryotic cells that store information not only in synapses, but also in their dendritic structure and connectivity, as well as genetic "marking" in the epigenome of each individual cell. Each neuron computes a complex nonlinear function of its inputs, roughly equivalent in processing capacity to an entire 1990s-era neural network model. Furthermore, individual cells provide the biological interface between gene expression, ongoing neural processing, and stored long-term memory traces. Neurons in all organisms have these properties, which are thus relevant to all of neuroscience and cognitive biology. Single-cell computation may also play a particular role in explaining some unusual features of human cognition. The recognition of the centrality of cellular computation to "natural computation" in brains, and of the constraints it imposes upon brain evolution, thus has important implications for the evolution of cognition, and how we study it.
Collapse
Affiliation(s)
- W. Tecumseh Fitch
- Faculty of Life Sciences and Vienna Cognitive Science Hub, University of Vienna, Vienna, Austria
| |
Collapse
|
37
|
Obara K, Ebina T, Terada SI, Uka T, Komatsu M, Takaji M, Watakabe A, Kobayashi K, Masamizu Y, Mizukami H, Yamamori T, Kasai K, Matsuzaki M. Change detection in the primate auditory cortex through feedback of prediction error signals. Nat Commun 2023; 14:6981. [PMID: 37957168 PMCID: PMC10643402 DOI: 10.1038/s41467-023-42553-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Accepted: 10/13/2023] [Indexed: 11/15/2023] Open
Abstract
Although cortical feedback signals are essential for modulating feedforward processing, no feedback error signal across hierarchical cortical areas has been reported. Here, we observed such a signal in the auditory cortex of awake common marmoset during an oddball paradigm to induce auditory duration mismatch negativity. Prediction errors to a deviant tone presentation were generated as offset calcium responses of layer 2/3 neurons in the rostral parabelt (RPB) of higher-order auditory cortex, while responses to non-deviant tones were strongly suppressed. Within several hundred milliseconds, the error signals propagated broadly into layer 1 of the primary auditory cortex (A1) and accumulated locally on top of incoming auditory signals. Blockade of RPB activity prevented deviance detection in A1. Optogenetic activation of RPB following tone presentation nonlinearly enhanced A1 tone response. Thus, the feedback error signal is critical for automatic detection of unpredicted stimuli in physiological auditory processing and may serve as backpropagation-like learning.
Collapse
Affiliation(s)
- Keitaro Obara
- Department of Physiology, Graduate School of Medicine, The University of Tokyo, Tokyo, 113-0033, Japan
- Brain Functional Dynamics Collaboration Laboratory, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
| | - Teppei Ebina
- Department of Physiology, Graduate School of Medicine, The University of Tokyo, Tokyo, 113-0033, Japan
| | - Shin-Ichiro Terada
- Department of Physiology, Graduate School of Medicine, The University of Tokyo, Tokyo, 113-0033, Japan
| | - Takanori Uka
- Department of Integrative Physiology, Graduate School of Medicine, University of Yamanashi, Yamanashi, 409-3898, Japan
| | - Misako Komatsu
- Laboratory for Molecular Analysis of Higher Brain Function, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
| | - Masafumi Takaji
- Laboratory for Molecular Analysis of Higher Brain Function, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
- Laboratory for Haptic Perception and Cognitive Physiology, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
| | - Akiya Watakabe
- Laboratory for Molecular Analysis of Higher Brain Function, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
- Laboratory for Haptic Perception and Cognitive Physiology, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
| | - Kenta Kobayashi
- Section of Viral Vector Development, National Institute for Physiological Sciences, Aichi, 444-8585, Japan
| | - Yoshito Masamizu
- Brain Functional Dynamics Collaboration Laboratory, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
| | - Hiroaki Mizukami
- Division of Genetic Therapeutics, Center for Molecular Medicine, Jichi Medical University, Tochigi, 329-0498, Japan
| | - Tetsuo Yamamori
- Laboratory for Molecular Analysis of Higher Brain Function, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
- Laboratory for Haptic Perception and Cognitive Physiology, RIKEN Center for Brain Science, Saitama, 351-0198, Japan
- Central Institute of Experimental Animals, Kanagawa, 210-0821, Japan
| | - Kiyoto Kasai
- Department of Neuropsychiatry, Graduate School of Medicine, The University of Tokyo, Tokyo, 113-8655, Japan
- International Research Center for Neurointelligence (WPI-IRCN), The University of Tokyo Institutes for Advanced Study, Tokyo, 113-0033, Japan
| | - Masanori Matsuzaki
- Department of Physiology, Graduate School of Medicine, The University of Tokyo, Tokyo, 113-0033, Japan.
- Brain Functional Dynamics Collaboration Laboratory, RIKEN Center for Brain Science, Saitama, 351-0198, Japan.
- International Research Center for Neurointelligence (WPI-IRCN), The University of Tokyo Institutes for Advanced Study, Tokyo, 113-0033, Japan.
| |
Collapse
|
38
|
Shtyrov Y, Efremov A, Kuptsova A, Wennekers T, Gutkin B, Garagnani M. Breakdown of category-specific word representations in a brain-constrained neurocomputational model of semantic dementia. Sci Rep 2023; 13:19572. [PMID: 37949997 PMCID: PMC10638411 DOI: 10.1038/s41598-023-41922-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 09/04/2023] [Indexed: 11/12/2023] Open
Abstract
The neurobiological nature of semantic knowledge, i.e., the encoding and storage of conceptual information in the human brain, remains a poorly understood and hotly debated subject. Clinical data on semantic deficits and neuroimaging evidence from healthy individuals have suggested multiple cortical regions to be involved in the processing of meaning. These include semantic hubs (most notably, anterior temporal lobe, ATL) that take part in semantic processing in general as well as sensorimotor areas that process specific aspects/categories according to their modality. Biologically inspired neurocomputational models can help elucidate the exact roles of these regions in the functioning of the semantic system and, importantly, in its breakdown in neurological deficits. We used a neuroanatomically constrained computational model of frontotemporal cortices implicated in word acquisition and processing, and adapted it to simulate and explain the effects of semantic dementia (SD) on word processing abilities. SD is a devastating, yet insufficiently understood progressive neurodegenerative disease, characterised by semantic knowledge deterioration that is hypothesised to be specifically related to neural damage in the ATL. The behaviour of our brain-based model is in full accordance with clinical data-namely, word comprehension performance decreases as SD lesions in ATL progress, whereas word repetition abilities remain less affected. Furthermore, our model makes predictions about lesion- and category-specific effects of SD: our simulation results indicate that word processing should be more impaired for object- than for action-related words, and that degradation of white matter should produce more severe consequences than the same proportion of grey matter decay. In sum, the present results provide a neuromechanistic explanatory account of cortical-level language impairments observed during the onset and progress of semantic dementia.
Collapse
Affiliation(s)
- Yury Shtyrov
- Center of Functionally Integrative Neuroscience (CFIN), Institute for Clinical Medicine, Aarhus University, Aarhus, Denmark.
| | - Aleksei Efremov
- Centre for Cognition and Decision Making, Institute for Cognitive Neuroscience, HSE University, Moscow, Russia
- Montreal Neurological Institute-Hospital, McGill University, Montreal, Quebec, Canada
| | - Anastasia Kuptsova
- Centre for Cognition and Decision Making, Institute for Cognitive Neuroscience, HSE University, Moscow, Russia
| | - Thomas Wennekers
- School of Engineering, Computing and Mathematics, University of Plymouth, Plymouth, UK
| | - Boris Gutkin
- Centre for Cognition and Decision Making, Institute for Cognitive Neuroscience, HSE University, Moscow, Russia
- Département d'Etudes Cognitives, École Normale Supérieure, Paris, France
| | - Max Garagnani
- Department of Computing, Goldsmiths - University of London, London, UK.
- Brain Language Laboratory, Department of Philosophy and Humanities, Freie Universität Berlin, Berlin, Germany.
| |
Collapse
|
39
|
Francioni V, Tang VD, Brown NJ, Toloza EH, Harnett M. Vectorized instructive signals in cortical dendrites during a brain-computer interface task. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.11.03.565534. [PMID: 37961227 PMCID: PMC10635122 DOI: 10.1101/2023.11.03.565534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/15/2023]
Abstract
Backpropagation of error is the most widely used learning algorithm in artificial neural networks, forming the backbone of modern machine learning and artificial intelligence1,2. Backpropagation provides a solution to the credit assignment problem by vectorizing an error signal tailored to individual neurons. Recent theoretical models have suggested that neural circuits could implement backpropagation-like learning by semi-independently processing feedforward and feedback information streams in separate dendritic compartments3-7. This presents a compelling, but untested, hypothesis for how cortical circuits could solve credit assignment in the brain. We designed a neurofeedback brain-computer interface (BCI) task with an experimenter-defined reward function to evaluate the key requirements for dendrites to implement backpropagation-like learning. We trained mice to modulate the activity of two spatially intermingled populations (4 or 5 neurons each) of layer 5 pyramidal neurons in the retrosplenial cortex to rotate a visual grating towards a target orientation while we recorded GCaMP activity from somas and corresponding distal apical dendrites. We observed that the relative magnitudes of somatic versus dendritic signals could be predicted using the activity of the surrounding network and contained information about task-related variables that could serve as instructive signals, including reward and error. The signs of these putative teaching signals both depended on the causal role of individual neurons in the task and predicted changes in overall activity over the course of learning. These results provide the first biological evidence of a backpropagation-like solution to the credit assignment problem in the brain.
Collapse
Affiliation(s)
- Valerio Francioni
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Vincent D Tang
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Norma J. Brown
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Enrique H.S. Toloza
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| | - Mark Harnett
- McGovern Institute for Brain Research, MIT, Cambridge, MA, USA
- Department of Brain and Cognitive Sciences, MIT, Cambridge, MA, USA
| |
Collapse
|
40
|
Pulvermüller F. Neurobiological mechanisms for language, symbols and concepts: Clues from brain-constrained deep neural networks. Prog Neurobiol 2023; 230:102511. [PMID: 37482195 PMCID: PMC10518464 DOI: 10.1016/j.pneurobio.2023.102511] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2022] [Revised: 05/02/2023] [Accepted: 07/18/2023] [Indexed: 07/25/2023]
Abstract
Neural networks are successfully used to imitate and model cognitive processes. However, to provide clues about the neurobiological mechanisms enabling human cognition, these models need to mimic the structure and function of real brains. Brain-constrained networks differ from classic neural networks by implementing brain similarities at different scales, ranging from the micro- and mesoscopic levels of neuronal function, local neuronal links and circuit interaction to large-scale anatomical structure and between-area connectivity. This review shows how brain-constrained neural networks can be applied to study in silico the formation of mechanisms for symbol and concept processing and to work towards neurobiological explanations of specifically human cognitive abilities. These include verbal working memory and learning of large vocabularies of symbols, semantic binding carried by specific areas of cortex, attention focusing and modulation driven by symbol type, and the acquisition of concrete and abstract concepts partly influenced by symbols. Neuronal assembly activity in the networks is analyzed to deliver putative mechanistic correlates of higher cognitive processes and to develop candidate explanations founded in established neurobiological principles.
Collapse
Affiliation(s)
- Friedemann Pulvermüller
- Brain Language Laboratory, Department of Philosophy and Humanities, WE4, Freie Universität Berlin, 14195 Berlin, Germany; Berlin School of Mind and Brain, Humboldt Universität zu Berlin, 10099 Berlin, Germany; Einstein Center for Neurosciences Berlin, 10117 Berlin, Germany; Cluster of Excellence 'Matters of Activity', Humboldt Universität zu Berlin, 10099 Berlin, Germany.
| |
Collapse
|
41
|
Halvagal MS, Zenke F. The combination of Hebbian and predictive plasticity learns invariant object representations in deep sensory networks. Nat Neurosci 2023; 26:1906-1915. [PMID: 37828226 PMCID: PMC10620089 DOI: 10.1038/s41593-023-01460-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Accepted: 09/08/2023] [Indexed: 10/14/2023]
Abstract
Recognition of objects from sensory stimuli is essential for survival. To that end, sensory networks in the brain must form object representations invariant to stimulus changes, such as size, orientation and context. Although Hebbian plasticity is known to shape sensory networks, it fails to create invariant object representations in computational models, raising the question of how the brain achieves such processing. In the present study, we show that combining Hebbian plasticity with a predictive form of plasticity leads to invariant representations in deep neural network models. We derive a local learning rule that generalizes to spiking neural networks and naturally accounts for several experimentally observed properties of synaptic plasticity, including metaplasticity and spike-timing-dependent plasticity. Finally, our model accurately captures neuronal selectivity changes observed in the primate inferotemporal cortex in response to altered visual experience. Thus, we provide a plausible normative theory emphasizing the importance of predictive plasticity mechanisms for successful representational learning.
Collapse
Affiliation(s)
- Manu Srinath Halvagal
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland
- Faculty of Science, University of Basel, Basel, Switzerland
| | - Friedemann Zenke
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland.
- Faculty of Science, University of Basel, Basel, Switzerland.
| |
Collapse
|
42
|
Amaya C, von Arnim A. Neurorobotic reinforcement learning for domains with parametrical uncertainty. Front Neurorobot 2023; 17:1239581. [PMID: 37965072 PMCID: PMC10642204 DOI: 10.3389/fnbot.2023.1239581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Accepted: 09/26/2023] [Indexed: 11/16/2023] Open
Abstract
Neuromorphic hardware paired with brain-inspired learning strategies have enormous potential for robot control. Explicitly, these advantages include low energy consumption, low latency, and adaptability. Therefore, developing and improving learning strategies, algorithms, and neuromorphic hardware integration in simulation is a key to moving the state-of-the-art forward. In this study, we used the neurorobotics platform (NRP) simulation framework to implement spiking reinforcement learning control for a robotic arm. We implemented a force-torque feedback-based classic object insertion task ("peg-in-hole") and controlled the robot for the first time with neuromorphic hardware in the loop. We therefore provide a solution for training the system in uncertain environmental domains by using randomized simulation parameters. This leads to policies that are robust to real-world parameter variations in the target domain, filling the sim-to-real gap.To the best of our knowledge, it is the first neuromorphic implementation of the peg-in-hole task in simulation with the neuromorphic Loihi chip in the loop, and with scripted accelerated interactive training in the Neurorobotics Platform, including randomized domains.
Collapse
Affiliation(s)
| | - Axel von Arnim
- Department of Neuromorphic Computing, Fortiss-Research Institute, Munich, Bavaria, Germany
| |
Collapse
|
43
|
Pham TQ, Matsui T, Chikazoe J. Evaluation of the Hierarchical Correspondence between the Human Brain and Artificial Neural Networks: A Review. BIOLOGY 2023; 12:1330. [PMID: 37887040 PMCID: PMC10604784 DOI: 10.3390/biology12101330] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Revised: 09/22/2023] [Accepted: 10/10/2023] [Indexed: 10/28/2023]
Abstract
Artificial neural networks (ANNs) that are heavily inspired by the human brain now achieve human-level performance across multiple task domains. ANNs have thus drawn attention in neuroscience, raising the possibility of providing a framework for understanding the information encoded in the human brain. However, the correspondence between ANNs and the brain cannot be measured directly. They differ in outputs and substrates, neurons vastly outnumber their ANN analogs (i.e., nodes), and the key algorithm responsible for most of modern ANN training (i.e., backpropagation) is likely absent from the brain. Neuroscientists have thus taken a variety of approaches to examine the similarity between the brain and ANNs at multiple levels of their information hierarchy. This review provides an overview of the currently available approaches and their limitations for evaluating brain-ANN correspondence.
Collapse
Affiliation(s)
| | - Teppei Matsui
- Graduate School of Brain Science, Doshisha University, Kyoto 610-0321, Japan
| | | |
Collapse
|
44
|
Konishi M, Igarashi KM, Miura K. Biologically plausible local synaptic learning rules robustly implement deep supervised learning. Front Neurosci 2023; 17:1160899. [PMID: 37886676 PMCID: PMC10598703 DOI: 10.3389/fnins.2023.1160899] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2023] [Accepted: 08/31/2023] [Indexed: 10/28/2023] Open
Abstract
In deep neural networks, representational learning in the middle layer is essential for achieving efficient learning. However, the currently prevailing backpropagation learning rules (BP) are not necessarily biologically plausible and cannot be implemented in the brain in their current form. Therefore, to elucidate the learning rules used by the brain, it is critical to establish biologically plausible learning rules for practical memory tasks. For example, learning rules that result in a learning performance worse than that of animals observed in experimental studies may not be computations used in real brains and should be ruled out. Using numerical simulations, we developed biologically plausible learning rules to solve a task that replicates a laboratory experiment where mice learned to predict the correct reward amount. Although the extreme learning machine (ELM) and weight perturbation (WP) learning rules performed worse than the mice, the feedback alignment (FA) rule achieved a performance equal to that of BP. To obtain a more biologically plausible model, we developed a variant of FA, FA_Ex-100%, which implements direct dopamine inputs that provide error signals locally in the layer of focus, as found in the mouse entorhinal cortex. The performance of FA_Ex-100% was comparable to that of conventional BP. Finally, we tested whether FA_Ex-100% was robust against rule perturbations and biologically inevitable noise. FA_Ex-100% worked even when subjected to perturbations, presumably because it could calibrate the correct prediction error (e.g., dopaminergic signals) in the next step as a teaching signal if the perturbation created a deviation. These results suggest that simplified and biologically plausible learning rules, such as FA_Ex-100%, can robustly facilitate deep supervised learning when the error signal, possibly conveyed by dopaminergic neurons, is accurate.
Collapse
Affiliation(s)
- Masataka Konishi
- Department of Biosciences, School of Biological and Environmental Sciences, Kwansei Gakuin University, Sanda, Hyogo, Japan
| | - Kei M. Igarashi
- Department of Anatomy and Neurobiology, School of Medicine, University of California, Irvine, Irvine, CA, United States
| | - Keiji Miura
- Department of Biosciences, School of Biological and Environmental Sciences, Kwansei Gakuin University, Sanda, Hyogo, Japan
| |
Collapse
|
45
|
Xu B, Poo MM. Large language models and brain-inspired general intelligence. Natl Sci Rev 2023; 10:nwad267. [PMID: 37942481 PMCID: PMC10630093 DOI: 10.1093/nsr/nwad267] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Indexed: 11/10/2023] Open
Affiliation(s)
- Bo Xu
- Institute of Automation, Chinese Academy of Sciences, China
| | - Mu-ming Poo
- CAS Center for Excellence for Brain Science and Intelligence Technology, China
| |
Collapse
|
46
|
Palaniyappan L, Benrimoh D, Voppel A, Rocca R. Studying Psychosis Using Natural Language Generation: A Review of Emerging Opportunities. BIOLOGICAL PSYCHIATRY. COGNITIVE NEUROSCIENCE AND NEUROIMAGING 2023; 8:994-1004. [PMID: 38441079 DOI: 10.1016/j.bpsc.2023.04.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/17/2023] [Revised: 04/16/2023] [Accepted: 04/19/2023] [Indexed: 03/07/2024]
Abstract
Disrupted language in psychotic disorders, such as schizophrenia, can manifest as false contents and formal deviations, often described as thought disorder. These features play a critical role in the social dysfunction associated with psychosis, but we continue to lack insights regarding how and why these symptoms develop. Natural language generation (NLG) is a field of computer science that focuses on generating human-like language for various applications. The theory that psychosis is related to the evolution of language in humans suggests that NLG systems that are sufficiently evolved to generate human-like language may also exhibit psychosis-like features. In this conceptual review, we propose using NLG systems that are at various stages of development as in silico tools to study linguistic features of psychosis. We argue that a program of in silico experimental research on the network architecture, function, learning rules, and training of NLG systems can help us understand better why thought disorder occurs in patients. This will allow us to gain a better understanding of the relationship between language and psychosis and potentially pave the way for new therapeutic approaches to address this vexing challenge.
Collapse
Affiliation(s)
- Lena Palaniyappan
- Douglas Mental Health University Institute, Department of Psychiatry, McGill University, Montreal, Quebec, Canada; Robarts Research Institute, Western University, London, Ontario, Canada; Department of Medical Biophysics, Western University, London, Ontario, Canada.
| | - David Benrimoh
- Douglas Mental Health University Institute, Department of Psychiatry, McGill University, Montreal, Quebec, Canada; Department of Psychiatry, Stanford University, Palo Alto, California
| | - Alban Voppel
- Douglas Mental Health University Institute, Department of Psychiatry, McGill University, Montreal, Quebec, Canada; Department of Psychiatry, University of Groningen, Groningen, the Netherlands
| | - Roberta Rocca
- Interacting Minds Centre, Department of Culture, Cognition and Computation, Aarhus University, Aarhus, Denmark
| |
Collapse
|
47
|
Liu H, Qin Y, Chen HY, Wu J, Ma J, Du Z, Wang N, Zou J, Lin S, Zhang X, Zhang Y, Wang H. Artificial Neuronal Devices Based on Emerging Materials: Neuronal Dynamics and Applications. ADVANCED MATERIALS (DEERFIELD BEACH, FLA.) 2023; 35:e2205047. [PMID: 36609920 DOI: 10.1002/adma.202205047] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/03/2022] [Revised: 12/02/2022] [Indexed: 06/17/2023]
Abstract
Artificial neuronal devices are critical building blocks of neuromorphic computing systems and currently the subject of intense research motivated by application needs from new computing technology and more realistic brain emulation. Researchers have proposed a range of device concepts that can mimic neuronal dynamics and functions. Although the switching physics and device structures of these artificial neurons are largely different, their behaviors can be described by several neuron models in a more unified manner. In this paper, the reports of artificial neuronal devices based on emerging volatile switching materials are reviewed from the perspective of the demonstrated neuron models, with a focus on the neuronal functions implemented in these devices and the exploitation of these functions for computational and sensing applications. Furthermore, the neuroscience inspirations and engineering methods to enrich the neuronal dynamics that remain to be implemented in artificial neuronal devices and networks toward realizing the full functionalities of biological neurons are discussed.
Collapse
Affiliation(s)
- Hefei Liu
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Yuan Qin
- Center for Power Electronics Systems, Bradley Department of Electrical and Computer Engineering, Virginia Polytechnic Institute and State University, Blacksburg, VA, 24060, USA
| | - Hung-Yu Chen
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jiangbin Wu
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jiahui Ma
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Zhonghao Du
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
| | - Nan Wang
- Mork Family Department of Chemical Engineering and Materials Science, University of Southern California, Los Angeles, CA, 90089, USA
| | - Jingyi Zou
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Sen Lin
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Xu Zhang
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
| | - Yuhao Zhang
- Center for Power Electronics Systems, Bradley Department of Electrical and Computer Engineering, Virginia Polytechnic Institute and State University, Blacksburg, VA, 24060, USA
| | - Han Wang
- Ming Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, 90089, USA
- Mork Family Department of Chemical Engineering and Materials Science, University of Southern California, Los Angeles, CA, 90089, USA
| |
Collapse
|
48
|
Paroli B, Martini G, Potenza MAC, Siano M, Mirigliano M, Milani P. Solving classification tasks by a receptron based on nonlinear optical speckle fields. Neural Netw 2023; 166:634-644. [PMID: 37604074 DOI: 10.1016/j.neunet.2023.08.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Revised: 06/07/2023] [Accepted: 08/02/2023] [Indexed: 08/23/2023]
Abstract
Among several approaches to tackle the problem of energy consumption in modern computing systems, two solutions are currently investigated: one consists of artificial neural networks (ANNs) based on photonic technologies, the other is a different paradigm compared to ANNs and it is based on random networks of non-linear nanoscale junctions resulting from the assembling of nanoparticles or nanowires as substrates for neuromorphic computing. These networks show the presence of emergent complexity and collective phenomena in analogy with biological neural networks characterized by self-organization, redundancy, and non-linearity. Starting from this background, we propose and formalize a generalization of the perceptron model to describe a classification device based on a network of interacting units where the input weights are non-linearly dependent. We show that this model, called "receptron", provides substantial advantages compared to the perceptron as, for example, the solution of non-linearly separable Boolean functions with a single device. The receptron model is used as a starting point for the implementation of an all-optical device that exploits the non-linearity of optical speckle fields produced by a solid scatterer. By encoding these speckle fields we generated a large variety of target Boolean functions. We demonstrate that by properly setting the model parameters, different classes of functions with different multiplicity can be solved efficiently. The optical implementation of the receptron scheme opens the way for the fabrication of a completely new class of optical devices for neuromorphic data processing based on a very simple hardware.
Collapse
Affiliation(s)
- B Paroli
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| | - G Martini
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| | - M A C Potenza
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| | - M Siano
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| | - M Mirigliano
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| | - P Milani
- CIMAINA and Dipartimento di Fisica, Università degli Studi di Milano, via G. Celoria 16, 20133, Milan, Italy.
| |
Collapse
|
49
|
Benjamin AS, Kording KP. A role for cortical interneurons as adversarial discriminators. PLoS Comput Biol 2023; 19:e1011484. [PMID: 37768890 PMCID: PMC10538760 DOI: 10.1371/journal.pcbi.1011484] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Accepted: 08/31/2023] [Indexed: 09/30/2023] Open
Abstract
The brain learns representations of sensory information from experience, but the algorithms by which it does so remain unknown. One popular theory formalizes representations as inferred factors in a generative model of sensory stimuli, meaning that learning must improve this generative model and inference procedure. This framework underlies many classic computational theories of sensory learning, such as Boltzmann machines, the Wake/Sleep algorithm, and a more recent proposal that the brain learns with an adversarial algorithm that compares waking and dreaming activity. However, in order for such theories to provide insights into the cellular mechanisms of sensory learning, they must be first linked to the cell types in the brain that mediate them. In this study, we examine whether a subtype of cortical interneurons might mediate sensory learning by serving as discriminators, a crucial component in an adversarial algorithm for representation learning. We describe how such interneurons would be characterized by a plasticity rule that switches from Hebbian plasticity during waking states to anti-Hebbian plasticity in dreaming states. Evaluating the computational advantages and disadvantages of this algorithm, we find that it excels at learning representations in networks with recurrent connections but scales poorly with network size. This limitation can be partially addressed if the network also oscillates between evoked activity and generative samples on faster timescales. Consequently, we propose that an adversarial algorithm with interneurons as discriminators is a plausible and testable strategy for sensory learning in biological systems.
Collapse
Affiliation(s)
- Ari S. Benjamin
- Department of Bioengineering, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
| | - Konrad P. Kording
- Department of Bioengineering, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
| |
Collapse
|
50
|
Tabatabaian F, Vora SR, Mirabbasi S. Applications, functions, and accuracy of artificial intelligence in restorative dentistry: A literature review. J ESTHET RESTOR DENT 2023; 35:842-859. [PMID: 37522291 DOI: 10.1111/jerd.13079] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Revised: 06/18/2023] [Accepted: 06/19/2023] [Indexed: 08/01/2023]
Abstract
OBJECTIVE The applications of artificial intelligence (AI) are increasing in restorative dentistry; however, the AI performance is unclear for dental professionals. The purpose of this narrative review was to evaluate the applications, functions, and accuracy of AI in diverse aspects of restorative dentistry including caries detection, tooth preparation margin detection, tooth restoration design, metal structure casting, dental restoration/implant detection, removable partial denture design, and tooth shade determination. OVERVIEW An electronic search was performed on Medline/PubMed, Embase, Web of Science, Cochrane, Scopus, and Google Scholar databases. English-language articles, published from January 1, 2000, to March 1, 2022, relevant to the aforementioned aspects were selected using the key terms of artificial intelligence, machine learning, deep learning, artificial neural networks, convolutional neural networks, clustering, soft computing, automated planning, computational learning, computer vision, and automated reasoning as inclusion criteria. A manual search was also performed. Therefore, 157 articles were included, reviewed, and discussed. CONCLUSIONS Based on the current literature, the AI models have shown promising performance in the mentioned aspects when being compared with traditional approaches in terms of accuracy; however, as these models are still in development, more studies are required to validate their accuracy and apply them to routine clinical practice. CLINICAL SIGNIFICANCE AI with its specific functions has shown successful applications with acceptable accuracy in diverse aspects of restorative dentistry. The understanding of these functions may lead to novel applications with optimal accuracy for AI in restorative dentistry.
Collapse
Affiliation(s)
- Farhad Tabatabaian
- Department of Oral Health Sciences, Faculty of Dentistry, The University of British Columbia, Vancouver, British Columbia, Canada
| | - Siddharth R Vora
- Department of Oral Health Sciences, Faculty of Dentistry, The University of British Columbia, Vancouver, British Columbia, Canada
| | - Shahriar Mirabbasi
- Department of Electrical and Computer Engineering, Faculty of Applied Science, The University of British Columbia, Vancouver, British Columbia, Canada
| |
Collapse
|