1
|
Saponati M, Vinck M. Sequence anticipation and spike-timing-dependent plasticity emerge from a predictive learning rule. Nat Commun 2023; 14:4985. [PMID: 37604825 PMCID: PMC10442404 DOI: 10.1038/s41467-023-40651-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2021] [Accepted: 08/03/2023] [Indexed: 08/23/2023] Open
Abstract
Intelligent behavior depends on the brain's ability to anticipate future events. However, the learning rules that enable neurons to predict and fire ahead of sensory inputs remain largely unknown. We propose a plasticity rule based on predictive processing, where the neuron learns a low-rank model of the synaptic input dynamics in its membrane potential. Neurons thereby amplify those synapses that maximally predict other synaptic inputs based on their temporal relations, which provide a solution to an optimization problem that can be implemented at the single-neuron level using only local information. Consequently, neurons learn sequences over long timescales and shift their spikes towards the first inputs in a sequence. We show that this mechanism can explain the development of anticipatory signalling and recall in a recurrent network. Furthermore, we demonstrate that the learning rule gives rise to several experimentally observed STDP (spike-timing-dependent plasticity) mechanisms. These findings suggest prediction as a guiding principle to orchestrate learning and synaptic plasticity in single neurons.
Collapse
Affiliation(s)
- Matteo Saponati
- Ernst Strüngmann Institute (ESI) for Neuroscience in Cooperation with Max Planck Society, 60528, Frankfurt Am Main, Germany.
- IMPRS for Neural Circuits, Max-Planck Institute for Brain Research, 60438, Frankfurt Am Main, Germany.
- Donders Centre for Neuroscience, Department of Neuroinformatics, Radboud University, 6525, Nijmegen, The Netherlands.
| | - Martin Vinck
- Ernst Strüngmann Institute (ESI) for Neuroscience in Cooperation with Max Planck Society, 60528, Frankfurt Am Main, Germany.
- Donders Centre for Neuroscience, Department of Neuroinformatics, Radboud University, 6525, Nijmegen, The Netherlands.
| |
Collapse
|
2
|
Yoon HG, Kim P. STDP-based associative memory formation and retrieval. J Math Biol 2023; 86:49. [PMID: 36826758 DOI: 10.1007/s00285-023-01883-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2021] [Revised: 12/11/2022] [Accepted: 01/31/2023] [Indexed: 02/25/2023]
Abstract
Spike-timing-dependent plasticity (STDP) is a biological process in which the precise order and timing of neuronal spikes affect the degree of synaptic modification. While there has been numerous research focusing on the role of STDP in neural coding, the functional implications of STDP at the macroscopic level in the brain have not been fully explored yet. In this work, we propose a neurodynamical model based on STDP that renders storage and retrieval of a group of associative memories. We showed that the function of STDP at the macroscopic level is to form a "memory plane" in the neural state space which dynamically encodes high dimensional data. We derived the analytic relation between the input, the memory plane, and the induced macroscopic neural oscillations around the memory plane. Such plane produces a limit cycle in reaction to a similar memory cue, which can be used for retrieval of the original input.
Collapse
Affiliation(s)
- Hong-Gyu Yoon
- Department of Mathematical Sciences, Ulsan National Institute of Science and Technology (UNIST), Ulsan Metropolitan City, 44919, Republic of Korea
| | - Pilwon Kim
- Department of Mathematical Sciences, Ulsan National Institute of Science and Technology (UNIST), Ulsan Metropolitan City, 44919, Republic of Korea.
| |
Collapse
|
3
|
Zhou W, Wen S, Liu Y, Liu L, Liu X, Chen L. Forgetting memristor based STDP learning circuit for neural networks. Neural Netw 2023; 158:293-304. [PMID: 36493532 DOI: 10.1016/j.neunet.2022.11.023] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2022] [Revised: 10/18/2022] [Accepted: 11/14/2022] [Indexed: 11/21/2022]
Abstract
The circuit implementation of STDP based on memristor is of great significance for the application of neural network. However, recent research shows that the research on the pure circuit implementation of forgetting memristor and STDP is still rare. This paper proposes a new STDP learning rule implementation circuit based on the forgetting memristor. This kind of forgetting memory resistance synapse makes the neural network have the function of time-division multiplexing, but the instability of short-term memory will affect the learning ability of the neural network. This paper analyzes and discusses the influence of synapses with long-term and short-term memory on the learning characteristics of neural network STDP, which lays a foundation for the construction of time-division multiplexing neural network with long-term and short-term memory synapses. Through this circuit, it is found that the volatile memristor has different behaviors to the stimulus signal in different initial states, and the resulting LTP phenomenon is more in line with the forgetting effect in biology. This circuit has multiple adjustable parameters, which can fit the STDP learning rules under different conditions. The application of neural network proves the availability of this circuit.
Collapse
Affiliation(s)
- Wenhao Zhou
- Electronic Information and Engineering, Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, 400715, China.
| | - Shiping Wen
- Centre for Artificial Intelligence, Faculty of Engineering and Information Technology, University of Technology Sydney, Australia.
| | - Yi Liu
- Electronic Information and Engineering, Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, 400715, China
| | - Lu Liu
- Electronic Information and Engineering, Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, 400715, China
| | - Xin Liu
- Computer Vision and Pattern Recognition Laboratory, School of Engineering Science, Lappeenranta-Lahti University of Technology LUT, Finland.
| | - Ling Chen
- Electronic Information and Engineering, Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, 400715, China; Computer Vision and Pattern Recognition Laboratory, School of Engineering Science, Lappeenranta-Lahti University of Technology LUT, Finland.
| |
Collapse
|
4
|
Kearney BE, Lanius RA. The brain-body disconnect: A somatic sensory basis for trauma-related disorders. Front Neurosci 2022; 16:1015749. [PMID: 36478879 PMCID: PMC9720153 DOI: 10.3389/fnins.2022.1015749] [Citation(s) in RCA: 27] [Impact Index Per Article: 13.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Accepted: 10/14/2022] [Indexed: 08/16/2023] Open
Abstract
Although the manifestation of trauma in the body is a phenomenon well-endorsed by clinicians and traumatized individuals, the neurobiological underpinnings of this manifestation remain unclear. The notion of somatic sensory processing, which encompasses vestibular and somatosensory processing and relates to the sensory systems concerned with how the physical body exists in and relates to physical space, is introduced as a major contributor to overall regulatory, social-emotional, and self-referential functioning. From a phylogenetically and ontogenetically informed perspective, trauma-related symptomology is conceptualized to be grounded in brainstem-level somatic sensory processing dysfunction and its cascading influences on physiological arousal modulation, affect regulation, and higher-order capacities. Lastly, we introduce a novel hierarchical model bridging somatic sensory processes with limbic and neocortical mechanisms regulating an individual's emotional experience and sense of a relational, agentive self. This model provides a working framework for the neurobiologically informed assessment and treatment of trauma-related conditions from a somatic sensory processing perspective.
Collapse
Affiliation(s)
- Breanne E. Kearney
- Department of Neuroscience, Schulich School of Medicine and Dentistry, Western University, London, ON, Canada
| | - Ruth A. Lanius
- Department of Neuroscience, Schulich School of Medicine and Dentistry, Western University, London, ON, Canada
- Department of Psychiatry, Schulich School of Medicine and Dentistry, Western University, London, ON, Canada
| |
Collapse
|
5
|
Yoon HG, Kim P. An STDP-based encoding method for associative and composite data. Sci Rep 2022; 12:4666. [PMID: 35304537 PMCID: PMC8933433 DOI: 10.1038/s41598-022-08469-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2021] [Accepted: 03/04/2022] [Indexed: 11/09/2022] Open
Abstract
Spike-timing-dependent plasticity(STDP) is a biological process of synaptic modification caused by the difference of firing order and timing between neurons. One of neurodynamical roles of STDP is to form a macroscopic geometrical structure in the neuronal state space in response to a periodic input by Susman et al. (Nat. Commun. 10(1), 1-9 2019), Yoon, & Kim. Stdp-based associative memory formation and retrieval. arXiv:2107.02429v2 (2021). In this work, we propose a practical memory model based on STDP which can store and retrieve high dimensional associative data. The model combines STDP dynamics with an encoding scheme for distributed representations and is able to handle multiple composite data in a continuous manner. In the auto-associative memory task where a group of images are continuously streamed to the model, the images are successfully retrieved from an oscillating neural state whenever a proper cue is given. In the second task that deals with semantic memories embedded from sentences, the results show that words can recall multiple sentences simultaneously or one exclusively, depending on their grammatical relations.
Collapse
Affiliation(s)
- Hong-Gyu Yoon
- Department of Mathematical Sciences, Ulsan National Institute of Science and Technology (UNIST), Ulsan Metropolitan City, 44919, Republic of Korea
| | - Pilwon Kim
- Department of Mathematical Sciences, Ulsan National Institute of Science and Technology (UNIST), Ulsan Metropolitan City, 44919, Republic of Korea.
| |
Collapse
|
6
|
Barnes J, Blair MR, Walshe RC, Tupper PF. LAG-1: A dynamic, integrative model of learning, attention, and gaze. PLoS One 2022; 17:e0259511. [PMID: 35298465 PMCID: PMC8929614 DOI: 10.1371/journal.pone.0259511] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2021] [Accepted: 10/21/2021] [Indexed: 11/19/2022] Open
Abstract
It is clear that learning and attention interact, but it is an ongoing challenge to integrate their psychological and neurophysiological descriptions. Here we introduce LAG-1, a dynamic neural field model of learning, attention and gaze, that we fit to human learning and eye-movement data from two category learning experiments. LAG-1 comprises three control systems: one for visuospatial attention, one for saccadic timing and control, and one for category learning. The model is able to extract a kind of information gain from pairwise differences in simple associations between visual features and categories. Providing this gain as a reentrant signal with bottom-up visual information, and in top-down spatial priority, appropriately influences the initiation of saccades. LAG-1 provides a moment-by-moment simulation of the interactions of learning and gaze, and thus simultaneously produces phenomena on many timescales, from the duration of saccades and gaze fixations, to the response times for trials, to the slow optimization of attention toward task relevant information across a whole experiment. With only three free parameters (learning rate, trial impatience, and fixation impatience) LAG-1 produces qualitatively correct fits for learning, behavioural timing and eye movement measures, and also for previously unmodelled empirical phenomena (e.g., fixation orders showing stimulus-specific attention, and decreasing fixation counts during feedback). Because LAG-1 is built to capture attention and gaze generally, we demonstrate how it can be applied to other phenomena of visual cognition such as the free viewing of visual stimuli, visual search, and covert attention.
Collapse
Affiliation(s)
- Jordan Barnes
- Department of Psychology, Simon Fraser University, Burnaby, BC, Canada
| | - Mark R. Blair
- Department of Psychology, Simon Fraser University, Burnaby, BC, Canada
- * E-mail:
| | - R. Calen Walshe
- Center for Perceptual Systems, University of Texas, Austin, Texas, United States of America
| | - Paul F. Tupper
- Department of Mathematics, Simon Fraser University, Burnaby, BC, Canada
| |
Collapse
|
7
|
Zhang K, Xu G, Du C, Liang R, Han C, Zheng X, Zhang S, Wang J, Tian P, Jia Y. Enhancement of capability for motor imagery using vestibular imbalance stimulation during brain computer interface. J Neural Eng 2021; 18. [PMID: 34571497 DOI: 10.1088/1741-2552/ac2a6f] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 09/27/2021] [Indexed: 01/07/2023]
Abstract
Objective.Motor imagery (MI), based on the theory of mirror neurons and neuroplasticity, can promote motor cortical activation in neurorehabilitation. The strategy of MI based on brain-computer interface (BCI) has been used in rehabilitation training and daily assistance for patients with hemiplegia in recent years. However, it is difficult to maintain the consistency and timeliness of receiving external stimulation to neural activation in most subjects owing to the high variability of electroencephalogram (EEG) representation across trials/subjects. Moreover, in practical application, MI-BCI cannot highly activate the motor cortex and provide stable interaction owing to the weakness of the EEG feature and lack of an effective mode of activation.Approach.In this study, a novel hybrid BCI paradigm based on MI and vestibular stimulation motor imagery (VSMI) was proposed to enhance the capability of feature response for MI. Twelve subjects participated in a group of controlled experiments containing VSMI and MI. Three indicators, namely, activation degree, timeliness, and classification accuracy, were adopted to evaluate the performance of the task.Main results.Vestibular stimulation could significantly strengthen the suppression ofαandβbands of contralateral brain regions during MI, that is, enhance the activation degree of the motor cortex (p< 0.01). Compared with MI, the timeliness of EEG feature-response achieved obvious improvements in VSMI experiments. Moreover, the averaged classification accuracy of VSMI and MI was 80.56% and 69.38%, respectively.Significance.The experimental results indicate that specific vestibular activity contributes to the oscillations of the motor cortex and has a positive effect on spontaneous imagery, which provides a novel MI paradigm and enables the preliminary exploration of sensorimotor integration of MI.
Collapse
Affiliation(s)
- Kai Zhang
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Guanghua Xu
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China.,State Key Laboratory for Manufacturing Systems Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Chenghang Du
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Renghao Liang
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Chenchen Han
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Xiaowei Zheng
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Sicong Zhang
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Jiahuan Wang
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Peiyuan Tian
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| | - Yaguang Jia
- School of Mechanical Engineering, Xi'an Jiaotong University, Xi'an, People's Republic of China
| |
Collapse
|
8
|
Real-time detection of bursts in neuronal cultures using a neuromorphic auditory sensor and spiking neural networks. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.03.109] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
9
|
Jia X, Hong H, DiCarlo JJ. Unsupervised changes in core object recognition behavior are predicted by neural plasticity in inferior temporal cortex. eLife 2021; 10:e60830. [PMID: 34114566 PMCID: PMC8324291 DOI: 10.7554/elife.60830] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2020] [Accepted: 06/10/2021] [Indexed: 11/13/2022] Open
Abstract
Temporal continuity of object identity is a feature of natural visual input and is potentially exploited - in an unsupervised manner - by the ventral visual stream to build the neural representation in inferior temporal (IT) cortex. Here, we investigated whether plasticity of individual IT neurons underlies human core object recognition behavioral changes induced with unsupervised visual experience. We built a single-neuron plasticity model combined with a previously established IT population-to-recognition-behavior-linking model to predict human learning effects. We found that our model, after constrained by neurophysiological data, largely predicted the mean direction, magnitude, and time course of human performance changes. We also found a previously unreported dependency of the observed human performance change on the initial task difficulty. This result adds support to the hypothesis that tolerant core object recognition in human and non-human primates is instructed - at least in part - by naturally occurring unsupervised temporal contiguity experience.
Collapse
Affiliation(s)
- Xiaoxuan Jia
- Department of Brain and Cognitive Sciences, Massachusetts Institute of TechnologyCambridgeUnited States
- McGovern Institute for Brain ResearchCambridgeUnited States
| | - Ha Hong
- Department of Brain and Cognitive Sciences, Massachusetts Institute of TechnologyCambridgeUnited States
- McGovern Institute for Brain ResearchCambridgeUnited States
- Harvard-MIT Division of Health Sciences and TechnologyCambridgeUnited States
| | - James J DiCarlo
- Department of Brain and Cognitive Sciences, Massachusetts Institute of TechnologyCambridgeUnited States
- McGovern Institute for Brain ResearchCambridgeUnited States
- Center for Brains, Minds and MachinesCambridgeUnited States
| |
Collapse
|
10
|
Abstract
Synaptic plasticity, the activity-dependent change in neuronal connection strength, has long been considered an important component of learning and memory. Computational and engineering work corroborate the power of learning through the directed adjustment of connection weights. Here we review the fundamental elements of four broadly categorized forms of synaptic plasticity and discuss their functional capabilities and limitations. Although standard, correlation-based, Hebbian synaptic plasticity has been the primary focus of neuroscientists for decades, it is inherently limited. Three-factor plasticity rules supplement Hebbian forms with neuromodulation and eligibility traces, while true supervised types go even further by adding objectives and instructive signals. Finally, a recently discovered hippocampal form of synaptic plasticity combines the above elements, while leaving behind the primary Hebbian requirement. We suggest that the effort to determine the neural basis of adaptive behavior could benefit from renewed experimental and theoretical investigation of more powerful directed types of synaptic plasticity.
Collapse
Affiliation(s)
- Jeffrey C Magee
- Department of Neuroscience and Howard Hughes Medical Institute, Baylor College of Medicine, Houston, Texas 77030, USA;
| | - Christine Grienberger
- Department of Neuroscience and Howard Hughes Medical Institute, Baylor College of Medicine, Houston, Texas 77030, USA;
| |
Collapse
|
11
|
Sinapayen L, Masumori A, Ikegami T. Reactive, Proactive, and Inductive Agents: An Evolutionary Path for Biological and Artificial Spiking Networks. Front Comput Neurosci 2020; 13:88. [PMID: 32038209 PMCID: PMC6987297 DOI: 10.3389/fncom.2019.00088] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2019] [Accepted: 12/16/2019] [Indexed: 11/29/2022] Open
Abstract
Complex environments provide structured yet variable sensory inputs. To best exploit information from these environments, organisms must evolve the ability to anticipate consequences of new stimuli, and act on these predictions. We propose an evolutionary path for neural networks, leading an organism from reactive behavior to simple proactive behavior and from simple proactive behavior to induction-based behavior. Based on earlier in-vitro and in-silico experiments, we define the conditions necessary in a network with spike-timing dependent plasticity for the organism to go from reactive to proactive behavior. Our results support the existence of specific evolutionary steps and four conditions necessary for embodied neural networks to evolve predictive and inductive abilities from an initial reactive strategy.
Collapse
Affiliation(s)
- Lana Sinapayen
- Sony Computer Science Laboratories, Inc., Tokyo, Japan.,Earth-Life Science Institute, Tokyo Institute of Technology, Tokyo, Japan
| | - Atsushi Masumori
- Department of General Systems Studies, Graduate School of Arts and Sciences, The University of Tokyo, Tokyo, Japan
| | - Takashi Ikegami
- Department of General Systems Studies, Graduate School of Arts and Sciences, The University of Tokyo, Tokyo, Japan
| |
Collapse
|
12
|
Towards spike-based machine intelligence with neuromorphic computing. Nature 2019; 575:607-617. [PMID: 31776490 DOI: 10.1038/s41586-019-1677-2] [Citation(s) in RCA: 329] [Impact Index Per Article: 65.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2018] [Accepted: 07/09/2019] [Indexed: 11/08/2022]
Abstract
Guided by brain-like 'spiking' computational frameworks, neuromorphic computing-brain-inspired computing for machine intelligence-promises to realize artificial intelligence while reducing the energy requirements of computing platforms. This interdisciplinary field began with the implementation of silicon circuits for biological neural routines, but has evolved to encompass the hardware implementation of algorithms with spike-based encoding and event-driven representations. Here we provide an overview of the developments in neuromorphic computing for both algorithms and hardware and highlight the fundamentals of learning and hardware frameworks. We discuss the main challenges and the future prospects of neuromorphic computing, with emphasis on algorithm-hardware codesign.
Collapse
|
13
|
Mikhael JG, Gershman SJ. Adapting the flow of time with dopamine. J Neurophysiol 2019; 121:1748-1760. [PMID: 30864882 PMCID: PMC6589719 DOI: 10.1152/jn.00817.2018] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2018] [Revised: 02/04/2019] [Accepted: 02/20/2019] [Indexed: 01/25/2023] Open
Abstract
The modulation of interval timing by dopamine (DA) has been well established over decades of research. The nature of this modulation, however, has remained controversial: Although the pharmacological evidence has largely suggested that time intervals are overestimated with higher DA levels, more recent optogenetic work has shown the opposite effect. In addition, a large body of work has asserted DA's role as a "reward prediction error" (RPE), or a teaching signal that allows the basal ganglia to learn to predict future rewards in reinforcement learning tasks. Whether these two seemingly disparate accounts of DA may be related has remained an open question. By taking a reinforcement learning-based approach to interval timing, we show here that the RPE interpretation of DA naturally extends to its role as a modulator of timekeeping and furthermore that this view reconciles the seemingly conflicting observations. We derive a biologically plausible, DA-dependent plasticity rule that can modulate the rate of timekeeping in either direction and whose effect depends on the timing of the DA signal itself. This bidirectional update rule can account for the results from pharmacology and optogenetics as well as the behavioral effects of reward rate on interval timing and the temporal selectivity of striatal neurons. Hence, by adopting a single RPE interpretation of DA, our results take a step toward unifying computational theories of reinforcement learning and interval timing. NEW & NOTEWORTHY How does dopamine (DA) influence interval timing? A large body of pharmacological evidence has suggested that DA accelerates timekeeping mechanisms. However, recent optogenetic work has shown exactly the opposite effect. In this article, we relate DA's role in timekeeping to its most established role, as a critical component of reinforcement learning. This allows us to derive a neurobiologically plausible framework that reconciles a large body of DA's temporal effects, including pharmacological, behavioral, electrophysiological, and optogenetic.
Collapse
Affiliation(s)
- John G Mikhael
- Program in Neuroscience and MD-PhD Program, Harvard Medical School , Boston, Massachusetts
| | - Samuel J Gershman
- Center for Brain Science and Department of Psychology, Harvard University , Cambridge, Massachusetts
| |
Collapse
|
14
|
Ferré P, Mamalet F, Thorpe SJ. Unsupervised Feature Learning With Winner-Takes-All Based STDP. Front Comput Neurosci 2018; 12:24. [PMID: 29674961 PMCID: PMC5895733 DOI: 10.3389/fncom.2018.00024] [Citation(s) in RCA: 39] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2017] [Accepted: 03/20/2018] [Indexed: 11/24/2022] Open
Abstract
We present a novel strategy for unsupervised feature learning in image applications inspired by the Spike-Timing-Dependent-Plasticity (STDP) biological learning rule. We show equivalence between rank order coding Leaky-Integrate-and-Fire neurons and ReLU artificial neurons when applied to non-temporal data. We apply this to images using rank-order coding, which allows us to perform a full network simulation with a single feed-forward pass using GPU hardware. Next we introduce a binary STDP learning rule compatible with training on batches of images. Two mechanisms to stabilize the training are also presented : a Winner-Takes-All (WTA) framework which selects the most relevant patches to learn from along the spatial dimensions, and a simple feature-wise normalization as homeostatic process. This learning process allows us to train multi-layer architectures of convolutional sparse features. We apply our method to extract features from the MNIST, ETH80, CIFAR-10, and STL-10 datasets and show that these features are relevant for classification. We finally compare these results with several other state of the art unsupervised learning methods.
Collapse
Affiliation(s)
- Paul Ferré
- Centre National de la Recherche Scientifique, UMR-5549, Toulouse, France.,Brainchip SAS, Balma, France
| | | | - Simon J Thorpe
- Centre National de la Recherche Scientifique, UMR-5549, Toulouse, France
| |
Collapse
|
15
|
Pauli WM, Cockburn J, Pool ER, Pérez OD, O’Doherty JP. Computational approaches to habits in a model-free world. Curr Opin Behav Sci 2018. [DOI: 10.1016/j.cobeha.2017.12.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
16
|
Min B, Zhou D, Cai D. Effects of Firing Variability on Network Structures with Spike-Timing-Dependent Plasticity. Front Comput Neurosci 2018; 12:1. [PMID: 29410621 PMCID: PMC5787127 DOI: 10.3389/fncom.2018.00001] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2017] [Accepted: 01/03/2018] [Indexed: 11/17/2022] Open
Abstract
Synaptic plasticity is believed to be the biological substrate underlying learning and memory. One of the most widespread forms of synaptic plasticity, spike-timing-dependent plasticity (STDP), uses the spike timing information of presynaptic and postsynaptic neurons to induce synaptic potentiation or depression. An open question is how STDP organizes the connectivity patterns in neuronal circuits. Previous studies have placed much emphasis on the role of firing rate in shaping connectivity patterns. Here, we go beyond the firing rate description to develop a self-consistent linear response theory that incorporates the information of both firing rate and firing variability. By decomposing the pairwise spike correlation into one component associated with local direct connections and the other associated with indirect connections, we identify two distinct regimes regarding the network structures learned through STDP. In one regime, the contribution of the direct-connection correlations dominates over that of the indirect-connection correlations in the learning dynamics; this gives rise to a network structure consistent with the firing rate description. In the other regime, the contribution of the indirect-connection correlations dominates in the learning dynamics, leading to a network structure different from the firing rate description. We demonstrate that the heterogeneity of firing variability across neuronal populations induces a temporally asymmetric structure of indirect-connection correlations. This temporally asymmetric structure underlies the emergence of the second regime. Our study provides a new perspective that emphasizes the role of high-order statistics of spiking activity in the spike-correlation-sensitive learning dynamics.
Collapse
Affiliation(s)
- Bin Min
- Center for Neural Science, Courant Institute of Mathematical Sciences, New York University, New York, NY, United States.,NYUAD Institute, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates
| | - Douglas Zhou
- School of Mathematical Sciences, MOE-LSC, Institute of Natural Sciences, Shanghai Jiao Tong University, Shanghai, China
| | - David Cai
- Center for Neural Science, Courant Institute of Mathematical Sciences, New York University, New York, NY, United States.,NYUAD Institute, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates.,School of Mathematical Sciences, MOE-LSC, Institute of Natural Sciences, Shanghai Jiao Tong University, Shanghai, China
| |
Collapse
|
17
|
Unsupervised heart-rate estimation in wearables with Liquid states and a probabilistic readout. Neural Netw 2018; 99:134-147. [PMID: 29414535 DOI: 10.1016/j.neunet.2017.12.015] [Citation(s) in RCA: 42] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2017] [Revised: 12/08/2017] [Accepted: 12/26/2017] [Indexed: 01/28/2023]
Abstract
Heart-rate estimation is a fundamental feature of modern wearable devices. In this paper we propose a machine learning technique to estimate heart-rate from electrocardiogram (ECG) data collected using wearable devices. The novelty of our approach lies in (1) encoding spatio-temporal properties of ECG signals directly into spike train and using this to excite recurrently connected spiking neurons in a Liquid State Machine computation model; (2) a novel learning algorithm; and (3) an intelligently designed unsupervised readout based on Fuzzy c-Means clustering of spike responses from a subset of neurons (Liquid states), selected using particle swarm optimization. Our approach differs from existing works by learning directly from ECG signals (allowing personalization), without requiring costly data annotations. Additionally, our approach can be easily implemented on state-of-the-art spiking-based neuromorphic systems, offering high accuracy, yet significantly low energy footprint, leading to an extended battery-life of wearable devices. We validated our approach with CARLsim, a GPU accelerated spiking neural network simulator modeling Izhikevich spiking neurons with Spike Timing Dependent Plasticity (STDP) and homeostatic scaling. A range of subjects is considered from in-house clinical trials and public ECG databases. Results show high accuracy and low energy footprint in heart-rate estimation across subjects with and without cardiac irregularities, signifying the strong potential of this approach to be integrated in future wearable devices.
Collapse
|
18
|
Russek EM, Momennejad I, Botvinick MM, Gershman SJ, Daw ND. Predictive representations can link model-based reinforcement learning to model-free mechanisms. PLoS Comput Biol 2017; 13:e1005768. [PMID: 28945743 PMCID: PMC5628940 DOI: 10.1371/journal.pcbi.1005768] [Citation(s) in RCA: 127] [Impact Index Per Article: 18.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2016] [Revised: 10/05/2017] [Accepted: 09/04/2017] [Indexed: 11/19/2022] Open
Abstract
Humans and animals are capable of evaluating actions by considering their long-run future rewards through a process described using model-based reinforcement learning (RL) algorithms. The mechanisms by which neural circuits perform the computations prescribed by model-based RL remain largely unknown; however, multiple lines of evidence suggest that neural circuits supporting model-based behavior are structurally homologous to and overlapping with those thought to carry out model-free temporal difference (TD) learning. Here, we lay out a family of approaches by which model-based computation may be built upon a core of TD learning. The foundation of this framework is the successor representation, a predictive state representation that, when combined with TD learning of value predictions, can produce a subset of the behaviors associated with model-based learning, while requiring less decision-time computation than dynamic programming. Using simulations, we delineate the precise behavioral capabilities enabled by evaluating actions using this approach, and compare them to those demonstrated by biological organisms. We then introduce two new algorithms that build upon the successor representation while progressively mitigating its limitations. Because this framework can account for the full range of observed putatively model-based behaviors while still utilizing a core TD framework, we suggest that it represents a neurally plausible family of mechanisms for model-based evaluation.
Collapse
Affiliation(s)
- Evan M. Russek
- Center for Neural Science, New York University, New York, NY, United States of America
| | - Ida Momennejad
- Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, United States of America
| | - Matthew M. Botvinick
- DeepMind, London, United Kingdom and Gatsby Computational Neuroscience Unit, University College London, United Kingdom
| | - Samuel J. Gershman
- Department of Psychology and Center for Brain Science, Harvard University, Cambridge, MA, United States of America
| | - Nathaniel D. Daw
- Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, United States of America
| |
Collapse
|
19
|
Budzillo A, Duffy A, Miller KE, Fairhall AL, Perkel DJ. Dopaminergic modulation of basal ganglia output through coupled excitation-inhibition. Proc Natl Acad Sci U S A 2017; 114:5713-5718. [PMID: 28507134 PMCID: PMC5465888 DOI: 10.1073/pnas.1611146114] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Learning and maintenance of skilled movements require exploration of motor space and selection of appropriate actions. Vocal learning and social context-dependent plasticity in songbirds depend on a basal ganglia circuit, which actively generates vocal variability. Dopamine in the basal ganglia reduces trial-to-trial neural variability when the bird engages in courtship song. Here, we present evidence for a unique, tonically active, excitatory interneuron in the songbird basal ganglia that makes strong synaptic connections onto output pallidal neurons, often linked in time with inhibitory events. Dopamine receptor activity modulates the coupling of these excitatory and inhibitory events in vitro, which results in a dynamic change in the synchrony of a modeled population of basal ganglia output neurons receiving excitatory and inhibitory inputs. The excitatory interneuron thus serves as one biophysical mechanism for the introduction or modulation of neural variability in this circuit.
Collapse
Affiliation(s)
- Agata Budzillo
- Graduate Program in Neurobiology and Behavior, University of Washington, Seattle, WA 98195
| | - Alison Duffy
- Department of Physics, University of Washington, Seattle, WA 98195
| | - Kimberly E Miller
- Department of Otolaryngology, University of Washington, Seattle, WA 98195
| | - Adrienne L Fairhall
- Department of Physiology and Biophysics, University of Washington, Seattle, WA 98195
- University of Washington Institute for Neuroengineering, University of Washington, Seattle, WA 98195
- Center for Sensorimotor Neural Engineering, University of Washington, Seattle, WA 98195
| | - David J Perkel
- Department of Otolaryngology, University of Washington, Seattle, WA 98195;
- University of Washington Institute for Neuroengineering, University of Washington, Seattle, WA 98195
- Department of Biology, University of Washington, Seattle, WA 98195
| |
Collapse
|
20
|
Guo L, Wang Z, Cabrerizo M, Adjouadi M. A Cross-Correlated Delay Shift Supervised Learning Method for Spiking Neurons with Application to Interictal Spike Detection in Epilepsy. Int J Neural Syst 2017; 27:1750002. [DOI: 10.1142/s0129065717500022] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
This study introduces a novel learning algorithm for spiking neurons, called CCDS, which is able to learn and reproduce arbitrary spike patterns in a supervised fashion allowing the processing of spatiotemporal information encoded in the precise timing of spikes. Unlike the Remote Supervised Method (ReSuMe), synapse delays and axonal delays in CCDS are variants which are modulated together with weights during learning. The CCDS rule is both biologically plausible and computationally efficient. The properties of this learning rule are investigated extensively through experimental evaluations in terms of reliability, adaptive learning performance, generality to different neuron models, learning in the presence of noise, effects of its learning parameters and classification performance. Results presented show that the CCDS learning method achieves learning accuracy and learning speed comparable with ReSuMe, but improves classification accuracy when compared to both the Spike Pattern Association Neuron (SPAN) learning rule and the Tempotron learning rule. The merit of CCDS rule is further validated on a practical example involving the automated detection of interictal spikes in EEG records of patients with epilepsy. Results again show that with proper encoding, the CCDS rule achieves good recognition performance.
Collapse
Affiliation(s)
- Lilin Guo
- Center for Advanced Technology and Education, Florida International University, 10555 W Flagler St, Miami, FL 33174, USA
| | - Zhenzhong Wang
- Center for Advanced Technology and Education, Florida International University, 10555 W Flagler St, Miami, FL 33174, USA
| | - Mercedes Cabrerizo
- Center for Advanced Technology and Education, Florida International University, 10555 W Flagler St, Miami, FL 33174, USA
| | - Malek Adjouadi
- Center for Advanced Technology and Education, Florida International University, 10555 W Flagler St, Miami, FL 33174, USA
| |
Collapse
|
21
|
Muller L, Piantoni G, Koller D, Cash SS, Halgren E, Sejnowski TJ. Rotating waves during human sleep spindles organize global patterns of activity that repeat precisely through the night. eLife 2016; 5:e17267. [PMID: 27855061 PMCID: PMC5114016 DOI: 10.7554/elife.17267] [Citation(s) in RCA: 91] [Impact Index Per Article: 11.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2016] [Accepted: 10/19/2016] [Indexed: 01/02/2023] Open
Abstract
During sleep, the thalamus generates a characteristic pattern of transient, 11-15 Hz sleep spindle oscillations, which synchronize the cortex through large-scale thalamocortical loops. Spindles have been increasingly demonstrated to be critical for sleep-dependent consolidation of memory, but the specific neural mechanism for this process remains unclear. We show here that cortical spindles are spatiotemporally organized into circular wave-like patterns, organizing neuronal activity over tens of milliseconds, within the timescale for storing memories in large-scale networks across the cortex via spike-time dependent plasticity. These circular patterns repeat over hours of sleep with millisecond temporal precision, allowing reinforcement of the activity patterns through hundreds of reverberations. These results provide a novel mechanistic account for how global sleep oscillations and synaptic plasticity could strengthen networks distributed across the cortex to store coherent and integrated memories.
Collapse
Affiliation(s)
- Lyle Muller
- Computational Neurobiology Laboratory, Salk Institute for Biological Studies, La Jolla, United States
| | - Giovanni Piantoni
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, United States
| | - Dominik Koller
- Computational Neurobiology Laboratory, Salk Institute for Biological Studies, La Jolla, United States
| | - Sydney S Cash
- Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, United States
| | - Eric Halgren
- Department of Radiology, University of California, San Diego, San Diego, United States
- Department of Neurosciences, University of California, San Diego, San Diego, United States
| | - Terrence J Sejnowski
- Computational Neurobiology Laboratory, Salk Institute for Biological Studies, La Jolla, United States
| |
Collapse
|
22
|
Abstract
To many, the poster child for David Marr's famous three levels of scientific inquiry is reinforcement learning-a computational theory of reward optimization, which readily prescribes algorithmic solutions that evidence striking resemblance to signals found in the brain, suggesting a straightforward neural implementation. Here we review questions that remain open at each level of analysis, concluding that the path forward to their resolution calls for inspiration across levels, rather than a focus on mutual constraints.
Collapse
Affiliation(s)
- Yael Niv
- Psychology Department & Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey, 08540
| | - Angela Langdon
- Psychology Department & Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey, 08540
| |
Collapse
|
23
|
Korte M, Schmitz D. Cellular and System Biology of Memory: Timing, Molecules, and Beyond. Physiol Rev 2016; 96:647-93. [PMID: 26960344 DOI: 10.1152/physrev.00010.2015] [Citation(s) in RCA: 73] [Impact Index Per Article: 9.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022] Open
Abstract
The storage of information in the mammalian nervous systems is dependent on a delicate balance between change and stability of neuronal networks. The induction and maintenance of processes that lead to changes in synaptic strength to a multistep process which can lead to long-lasting changes, which starts and ends with a highly choreographed and perfectly timed dance of molecules in different cell types of the central nervous system. This is accompanied by synchronization of specific networks, resulting in the generation of characteristic "macroscopic" rhythmic electrical fields, whose characteristic frequencies correspond to certain activity and information-processing states of the brain. Molecular events and macroscopic fields influence each other reciprocally. We review here cellular processes of synaptic plasticity, particularly functional and structural changes, and focus on timing events that are important for the initial memory acquisition, as well as mechanisms of short- and long-term memory storage. Then, we cover the importance of epigenetic events on the long-time range. Furthermore, we consider how brain rhythms at the network level participate in processes of information storage and by what means they participating in it. Finally, we examine memory consolidation at the system level during processes of sleep.
Collapse
Affiliation(s)
- Martin Korte
- Zoological Institute, Division of Cellular Neurobiology, Braunschweig, Germany; Helmholtz Centre for Infection Research, AG NIND, Braunschweig, Germany; and Neuroscience Research Centre, Charité Universitätsmedizin Berlin, Berlin, Germany
| | - Dietmar Schmitz
- Zoological Institute, Division of Cellular Neurobiology, Braunschweig, Germany; Helmholtz Centre for Infection Research, AG NIND, Braunschweig, Germany; and Neuroscience Research Centre, Charité Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
24
|
Bellec G, Galtier M, Brette R, Yger P. Slow feature analysis with spiking neurons and its application to audio stimuli. J Comput Neurosci 2016; 40:317-29. [PMID: 27075919 DOI: 10.1007/s10827-016-0599-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2015] [Revised: 01/31/2016] [Accepted: 02/29/2016] [Indexed: 10/22/2022]
Abstract
Extracting invariant features in an unsupervised manner is crucial to perform complex computation such as object recognition, analyzing music or understanding speech. While various algorithms have been proposed to perform such a task, Slow Feature Analysis (SFA) uses time as a means of detecting those invariants, extracting the slowly time-varying components in the input signals. In this work, we address the question of how such an algorithm can be implemented by neurons, and apply it in the context of audio stimuli. We propose a projected gradient implementation of SFA that can be adapted to a Hebbian like learning rule dealing with biologically plausible neuron models. Furthermore, we show that a Spike-Timing Dependent Plasticity learning rule, shaped as a smoothed second derivative, implements SFA for spiking neurons. The theory is supported by numerical simulations, and to illustrate a simple use of SFA, we have applied it to auditory signals. We show that a single SFA neuron can learn to extract the tempo in sound recordings.
Collapse
Affiliation(s)
- Guillaume Bellec
- Institut de la Vision, Sorbonne Université, UPMC Univ Paris06 UMRS968, Paris, France. .,INSERM, U968, Paris, France. .,CNRS, UMR7210, Paris, France.
| | - Mathieu Galtier
- European Institute for Theoretical Neuroscience CNRS UNIC UPR-3293, Paris, France
| | - Romain Brette
- Institut de la Vision, Sorbonne Université, UPMC Univ Paris06 UMRS968, Paris, France.,INSERM, U968, Paris, France.,CNRS, UMR7210, Paris, France
| | - Pierre Yger
- Institut de la Vision, Sorbonne Université, UPMC Univ Paris06 UMRS968, Paris, France.,INSERM, U968, Paris, France.,CNRS, UMR7210, Paris, France.,Institut d'Etudes de la Cognition, ENS, Paris, France
| |
Collapse
|
25
|
Veliz-Cuba A, Shouval HZ, Josić K, Kilpatrick ZP. Networks that learn the precise timing of event sequences. J Comput Neurosci 2015; 39:235-54. [PMID: 26334992 DOI: 10.1007/s10827-015-0574-4] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2015] [Revised: 08/06/2015] [Accepted: 08/10/2015] [Indexed: 11/28/2022]
Abstract
Neuronal circuits can learn and replay firing patterns evoked by sequences of sensory stimuli. After training, a brief cue can trigger a spatiotemporal pattern of neural activity similar to that evoked by a learned stimulus sequence. Network models show that such sequence learning can occur through the shaping of feedforward excitatory connectivity via long term plasticity. Previous models describe how event order can be learned, but they typically do not explain how precise timing can be recalled. We propose a mechanism for learning both the order and precise timing of event sequences. In our recurrent network model, long term plasticity leads to the learning of the sequence, while short term facilitation enables temporally precise replay of events. Learned synaptic weights between populations determine the time necessary for one population to activate another. Long term plasticity adjusts these weights so that the trained event times are matched during playback. While we chose short term facilitation as a time-tracking process, we also demonstrate that other mechanisms, such as spike rate adaptation, can fulfill this role. We also analyze the impact of trial-to-trial variability, showing how observational errors as well as neuronal noise result in variability in learned event times. The dynamics of the playback process determines how stochasticity is inherited in learned sequence timings. Future experiments that characterize such variability can therefore shed light on the neural mechanisms of sequence learning.
Collapse
Affiliation(s)
- Alan Veliz-Cuba
- Department of Mathematics, University of Houston, Houston, TX, 77204, USA.
| | - Harel Z Shouval
- Department of Neurobiology and Anatomy, University of Texas Medical School, Houston, TX, 77030, USA.
| | - Krešimir Josić
- Department of Mathematics, University of Houston, Houston, TX, 77204, USA. .,Department of Biology, University of Houston, Houston, TX, 77204, USA.
| | | |
Collapse
|
26
|
Fotouhi M, Heidari M, Sharifitabar M. Continuous neural network with windowed Hebbian learning. BIOLOGICAL CYBERNETICS 2015; 109:321-332. [PMID: 25677526 DOI: 10.1007/s00422-015-0645-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2013] [Accepted: 01/27/2015] [Indexed: 06/04/2023]
Abstract
We introduce an extension of the classical neural field equation where the dynamics of the synaptic kernel satisfies the standard Hebbian type of learning (synaptic plasticity). Here, a continuous network in which changes in the weight kernel occurs in a specified time window is considered. A novelty of this model is that it admits synaptic weight decrease as well as the usual weight increase resulting from correlated activity. The resulting equation leads to a delay-type rate model for which the existence and stability of solutions such as the rest state, bumps, and traveling fronts are investigated. Some relations between the length of the time window and the bump width is derived. In addition, the effect of the delay parameter on the stability of solutions is shown. Also numerical simulations for solutions and their stability are presented.
Collapse
Affiliation(s)
- M Fotouhi
- Department of Mathematical Sciences, Sharif University of Technology, P.O. Box 11365-9415, Tehran, Iran,
| | | | | |
Collapse
|
27
|
Skorheim S, Lonjers P, Bazhenov M. A spiking network model of decision making employing rewarded STDP. PLoS One 2014; 9:e90821. [PMID: 24632858 PMCID: PMC3954625 DOI: 10.1371/journal.pone.0090821] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2013] [Accepted: 02/05/2014] [Indexed: 01/08/2023] Open
Abstract
Reward-modulated spike timing dependent plasticity (STDP) combines unsupervised STDP with a reinforcement signal that modulates synaptic changes. It was proposed as a learning rule capable of solving the distal reward problem in reinforcement learning. Nonetheless, performance and limitations of this learning mechanism have yet to be tested for its ability to solve biological problems. In our work, rewarded STDP was implemented to model foraging behavior in a simulated environment. Over the course of training the network of spiking neurons developed the capability of producing highly successful decision-making. The network performance remained stable even after significant perturbations of synaptic structure. Rewarded STDP alone was insufficient to learn effective decision making due to the difficulty maintaining homeostatic equilibrium of synaptic weights and the development of local performance maxima. Our study predicts that successful learning requires stabilizing mechanisms that allow neurons to balance their input and output synapses as well as synaptic noise.
Collapse
Affiliation(s)
- Steven Skorheim
- Department of Cell Biology and Neuroscience, University of California Riverside, Riverside, California, United States of America
| | - Peter Lonjers
- Department of Cell Biology and Neuroscience, University of California Riverside, Riverside, California, United States of America
| | - Maxim Bazhenov
- Department of Cell Biology and Neuroscience, University of California Riverside, Riverside, California, United States of America
- * E-mail:
| |
Collapse
|
28
|
Krieg D, Triesch J. A unifying theory of synaptic long-term plasticity based on a sparse distribution of synaptic strength. Front Synaptic Neurosci 2014; 6:3. [PMID: 24624080 PMCID: PMC3941589 DOI: 10.3389/fnsyn.2014.00003] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2013] [Accepted: 02/13/2014] [Indexed: 11/30/2022] Open
Abstract
Long-term synaptic plasticity is fundamental to learning and network function. It has been studied under various induction protocols and depends on firing rates, membrane voltage, and precise timing of action potentials. These protocols show different facets of a common underlying mechanism but they are mostly modeled as distinct phenomena. Here, we show that all of these different dependencies can be explained from a single computational principle. The objective is a sparse distribution of excitatory synaptic strength, which may help to reduce metabolic costs associated with synaptic transmission. Based on this objective we derive a stochastic gradient ascent learning rule which is of differential-Hebbian type. It is formulated in biophysical quantities and can be related to current mechanistic theories of synaptic plasticity. The learning rule accounts for experimental findings from all major induction protocols and explains a classic phenomenon of metaplasticity. Furthermore, our model predicts the existence of metaplasticity for spike-timing-dependent plasticity Thus, we provide a theory of long-term synaptic plasticity that unifies different induction protocols and provides a connection between functional and mechanistic levels of description.
Collapse
Affiliation(s)
- Daniel Krieg
- Frankfurt Institute for Advanced Studies, Goethe University Frankfurt, Germany
| | - Jochen Triesch
- Frankfurt Institute for Advanced Studies, Goethe University Frankfurt, Germany
| |
Collapse
|
29
|
Kappel D, Nessler B, Maass W. STDP installs in Winner-Take-All circuits an online approximation to hidden Markov model learning. PLoS Comput Biol 2014; 10:e1003511. [PMID: 24675787 PMCID: PMC3967926 DOI: 10.1371/journal.pcbi.1003511] [Citation(s) in RCA: 48] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2013] [Accepted: 01/24/2014] [Indexed: 11/18/2022] Open
Abstract
In order to cross a street without being run over, we need to be able to extract very fast hidden causes of dynamically changing multi-modal sensory stimuli, and to predict their future evolution. We show here that a generic cortical microcircuit motif, pyramidal cells with lateral excitation and inhibition, provides the basis for this difficult but all-important information processing capability. This capability emerges in the presence of noise automatically through effects of STDP on connections between pyramidal cells in Winner-Take-All circuits with lateral excitation. In fact, one can show that these motifs endow cortical microcircuits with functional properties of a hidden Markov model, a generic model for solving such tasks through probabilistic inference. Whereas in engineering applications this model is adapted to specific tasks through offline learning, we show here that a major portion of the functionality of hidden Markov models arises already from online applications of STDP, without any supervision or rewards. We demonstrate the emergent computing capabilities of the model through several computer simulations. The full power of hidden Markov model learning can be attained through reward-gated STDP. This is due to the fact that these mechanisms enable a rejection sampling approximation to theoretically optimal learning. We investigate the possible performance gain that can be achieved with this more accurate learning method for an artificial grammar task.
Collapse
Affiliation(s)
- David Kappel
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| | - Bernhard Nessler
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| | - Wolfgang Maass
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| |
Collapse
|
30
|
Yger P, Harris KD. The Convallis rule for unsupervised learning in cortical networks. PLoS Comput Biol 2013; 9:e1003272. [PMID: 24204224 PMCID: PMC3808450 DOI: 10.1371/journal.pcbi.1003272] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2013] [Accepted: 08/28/2013] [Indexed: 01/26/2023] Open
Abstract
The phenomenology and cellular mechanisms of cortical synaptic plasticity are becoming known in increasing detail, but the computational principles by which cortical plasticity enables the development of sensory representations are unclear. Here we describe a framework for cortical synaptic plasticity termed the "Convallis rule", mathematically derived from a principle of unsupervised learning via constrained optimization. Implementation of the rule caused a recurrent cortex-like network of simulated spiking neurons to develop rate representations of real-world speech stimuli, enabling classification by a downstream linear decoder. Applied to spike patterns used in in vitro plasticity experiments, the rule reproduced multiple results including and beyond STDP. However STDP alone produced poorer learning performance. The mathematical form of the rule is consistent with a dual coincidence detector mechanism that has been suggested by experiments in several synaptic classes of juvenile neocortex. Based on this confluence of normative, phenomenological, and mechanistic evidence, we suggest that the rule may approximate a fundamental computational principle of the neocortex.
Collapse
Affiliation(s)
- Pierre Yger
- UCL Institute of Neurology and UCL Department of Neuroscience, Physiology, and Pharmacology, London, United Kingdom
- * E-mail:
| | - Kenneth D. Harris
- UCL Institute of Neurology and UCL Department of Neuroscience, Physiology, and Pharmacology, London, United Kingdom
| |
Collapse
|
31
|
Abstract
Numerous experimental data suggest that simultaneously or sequentially activated assemblies of neurons play a key role in the storage and computational use of long-term memory in the brain. However, a model that elucidates how these memory traces could emerge through spike-timing-dependent plasticity (STDP) has been missing. We show here that stimulus-specific assemblies of neurons emerge automatically through STDP in a simple cortical microcircuit model. The model that we consider is a randomly connected network of well known microcircuit motifs: pyramidal cells with lateral inhibition. We show that the emergent assembly codes for repeatedly occurring spatiotemporal input patterns tend to fire in some loose, sequential manner that is reminiscent of experimentally observed stereotypical trajectories of network states. We also show that the emergent assembly codes add an important computational capability to standard models for online computations in cortical microcircuits: the capability to integrate information from long-term memory with information from novel spike inputs.
Collapse
|
32
|
Vassiliades V, Christodoulou C. Toward nonlinear local reinforcement learning rules through neuroevolution. Neural Comput 2013; 25:3020-43. [PMID: 24001343 DOI: 10.1162/neco_a_00514] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
We consider the problem of designing local reinforcement learning rules for artificial neural network (ANN) controllers. Motivated by the universal approximation properties of ANNs, we adopt an ANN representation for the learning rules, which are optimized using evolutionary algorithms. We evaluate the ANN rules in partially observable versions of four tasks: the mountain car, the acrobot, the cart pole balancing, and the nonstationary mountain car. For testing whether such evolved ANN-based learning rules perform satisfactorily, we compare their performance with the performance of SARSA(λ) with tile coding, when the latter is provided with either full or partial state information. The comparison shows that the evolved rules perform much better than SARSA(λ) with partial state information and are comparable to the one with full state information, while in the case of the nonstationary environment, the evolved rule is much more adaptive. It is therefore clear that the proposed approach can be particularly effective in both partially observable and nonstationary environments. Moreover, it could potentially be utilized toward creating more general rules that can be applied in multiple domains and transfer learning scenarios.
Collapse
|
33
|
Galtier MN, Wainrib G. A biological gradient descent for prediction through a combination of STDP and homeostatic plasticity. Neural Comput 2013; 25:2815-32. [PMID: 24001342 DOI: 10.1162/neco_a_00512] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
Identifying, formalizing, and combining biological mechanisms that implement known brain functions, such as prediction, is a main aspect of research in theoretical neuroscience. In this letter, the mechanisms of spike-timing-dependent plasticity and homeostatic plasticity, combined in an original mathematical formalism, are shown to shape recurrent neural networks into predictors. Following a rigorous mathematical treatment, we prove that they implement the online gradient descent of a distance between the network activity and its stimuli. The convergence to an equilibrium, where the network can spontaneously reproduce or predict its stimuli, does not suffer from bifurcation issues usually encountered in learning in recurrent neural networks.
Collapse
Affiliation(s)
- Mathieu N Galtier
- School of Engineering and Science, Jacobs University Bremen gGmbH, 28759 Bremen, Germany
| | | |
Collapse
|
34
|
Abstract
In the absence of sensory input, neuronal networks are far from being silent. Whether spontaneous changes in ongoing activity reflect previous sensory experience or stochastic fluctuations in brain activity is not well understood. Here we describe reactivation of stimulus-evoked activity in awake visual cortical networks. We found that continuous exposure to randomly flashed image sequences induces reactivation in macaque V4 cortical networks in the absence of visual stimulation. This reactivation of previously evoked activity is stimulus-specific, occurs only in the same temporal order as the original response, and strengthens with increased stimulus exposures. Importantly, cells exhibiting significant reactivation carry more information about the stimulus than cells that do not reactivate. These results demonstrate a surprising degree of experience-dependent plasticity in visual cortical networks as a result of repeated exposure to unattended information. We suggest that awake reactivation in visual cortex may underlie perceptual learning by passive stimulus exposure.
Collapse
Affiliation(s)
- Sarah L. Eagleman
- Department of Neurobiology and Anatomy, University of Texas–Houston Medical School, Houston, TX 77030
| | - Valentin Dragoi
- Department of Neurobiology and Anatomy, University of Texas–Houston Medical School, Houston, TX 77030
| |
Collapse
|
35
|
What can neurons do for their brain? Communicate selectivity with bursts. Theory Biosci 2012; 132:27-39. [PMID: 22956291 DOI: 10.1007/s12064-012-0165-0] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2011] [Accepted: 07/30/2012] [Indexed: 10/27/2022]
Abstract
Neurons deep in cortex interact with the environment extremely indirectly; the spikes they receive and produce are pre- and post-processed by millions of other neurons. This paper proposes two information-theoretic constraints guiding the production of spikes, that help ensure bursting activity deep in cortex relates meaningfully to events in the environment. First, neurons should emphasize selective responses with bursts. Second, neurons should propagate selective inputs by burst-firing in response to them. We show the constraints are necessary for bursts to dominate information-transfer within cortex, thereby providing a substrate allowing neurons to distribute credit amongst themselves. Finally, since synaptic plasticity degrades the ability of neurons to burst selectively, we argue that homeostatic regulation of synaptic weights is necessary, and that it is best performed offline during sleep.
Collapse
|
36
|
Davies S, Galluppi F, Rast AD, Furber SB. A forecast-based STDP rule suitable for neuromorphic implementation. Neural Netw 2012; 32:3-14. [PMID: 22386500 DOI: 10.1016/j.neunet.2012.02.018] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2011] [Revised: 01/15/2012] [Accepted: 02/07/2012] [Indexed: 11/17/2022]
Abstract
Artificial neural networks increasingly involve spiking dynamics to permit greater computational efficiency. This becomes especially attractive for on-chip implementation using dedicated neuromorphic hardware. However, both spiking neural networks and neuromorphic hardware have historically found difficulties in implementing efficient, effective learning rules. The best-known spiking neural network learning paradigm is Spike Timing Dependent Plasticity (STDP) which adjusts the strength of a connection in response to the time difference between the pre- and post-synaptic spikes. Approaches that relate learning features to the membrane potential of the post-synaptic neuron have emerged as possible alternatives to the more common STDP rule, with various implementations and approximations. Here we use a new type of neuromorphic hardware, SpiNNaker, which represents the flexible "neuromimetic" architecture, to demonstrate a new approach to this problem. Based on the standard STDP algorithm with modifications and approximations, a new rule, called STDP TTS (Time-To-Spike) relates the membrane potential with the Long Term Potentiation (LTP) part of the basic STDP rule. Meanwhile, we use the standard STDP rule for the Long Term Depression (LTD) part of the algorithm. We show that on the basis of the membrane potential it is possible to make a statistical prediction of the time needed by the neuron to reach the threshold, and therefore the LTP part of the STDP algorithm can be triggered when the neuron receives a spike. In our system these approximations allow efficient memory access, reducing the overall computational time and the memory bandwidth required. The improvements here presented are significant for real-time applications such as the ones for which the SpiNNaker system has been designed. We present simulation results that show the efficacy of this algorithm using one or more input patterns repeated over the whole time of the simulation. On-chip results show that the STDP TTS algorithm allows the neural network to adapt and detect the incoming pattern with improvements both in the reliability of, and the time required for, consistent output. Through the approximations we suggest in this paper, we introduce a learning rule that is easy to implement both in event-driven simulators and in dedicated hardware, reducing computational complexity relative to the standard STDP rule. Such a rule offers a promising solution, complementary to standard STDP evaluation algorithms, for real-time learning using spiking neural networks in time-critical applications.
Collapse
Affiliation(s)
- S Davies
- School of Computer Science, The University of Manchester, Oxford Road, Manchester, M13 9PL, United Kingdom.
| | | | | | | |
Collapse
|
37
|
Burbank KS, Kreiman G. Depression-biased reverse plasticity rule is required for stable learning at top-down connections. PLoS Comput Biol 2012; 8:e1002393. [PMID: 22396630 PMCID: PMC3291526 DOI: 10.1371/journal.pcbi.1002393] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2011] [Accepted: 01/01/2012] [Indexed: 11/19/2022] Open
Abstract
Top-down synapses are ubiquitous throughout neocortex and play a central role in cognition, yet little is known about their development and specificity. During sensory experience, lower neocortical areas are activated before higher ones, causing top-down synapses to experience a preponderance of post-synaptic activity preceding pre-synaptic activity. This timing pattern is the opposite of that experienced by bottom-up synapses, which suggests that different versions of spike-timing dependent synaptic plasticity (STDP) rules may be required at top-down synapses. We consider a two-layer neural network model and investigate which STDP rules can lead to a distribution of top-down synaptic weights that is stable, diverse and avoids strong loops. We introduce a temporally reversed rule (rSTDP) where top-down synapses are potentiated if post-synaptic activity precedes pre-synaptic activity. Combining analytical work and integrate-and-fire simulations, we show that only depression-biased rSTDP (and not classical STDP) produces stable and diverse top-down weights. The conclusions did not change upon addition of homeostatic mechanisms, multiplicative STDP rules or weak external input to the top neurons. Our prediction for rSTDP at top-down synapses, which are distally located, is supported by recent neurophysiological evidence showing the existence of temporally reversed STDP in synapses that are distal to the post-synaptic cell body. The complex circuitry in the cerebral cortex is characterized by bottom-up connections, which carry feedforward information from the sensory periphery to higher areas, and top-down connections, where the information flow is reversed. Changes over time in the strength of synaptic connections between neurons underlie development, learning and memory. A fundamental mechanism to change synaptic strength is spike timing dependent plasticity, whereby synapses are strengthened whenever pre-synaptic spikes shortly precede post-synaptic spikes and are weakened otherwise; the relative timing of spikes therefore dictates the direction of plasticity. Spike timing dependent plasticity has been observed in multiple species and different brain areas. Here, we argue that top-down connections obey a learning rule with a reversed temporal dependence, which we call reverse spike timing dependent plasticity. We use mathematical analysis and computational simulations to show that this reverse time learning rule, and not previous learning rules, leads to a biologically plausible connectivity pattern with stable synaptic strengths. This reverse time learning rule is supported by recent neuroanatomical and neurophysiological experiments and can explain empirical observations about the development and function of top-down synapses in the brain.
Collapse
Affiliation(s)
- Kendra S. Burbank
- Department of Neurology and Ophthalmology, Children's Hospital Boston, Harvard Medical School, Boston, Massachusetts, United States of America
| | - Gabriel Kreiman
- Department of Neurology and Ophthalmology, Children's Hospital Boston, Harvard Medical School, Boston, Massachusetts, United States of America
- Center for Brain Science, Harvard University, Cambridge, Massachusetts, United States of America
- Swartz Center for Theoretical Neuroscience, Harvard University, Cambridge, Massachusetts, United States of America
- * E-mail:
| |
Collapse
|
38
|
BELATRECHE AMMAR, MAGUIRE LIAMP, MCGINNITY MARTIN, WU QINGXIANG. EVOLUTIONARY DESIGN OF SPIKING NEURAL NETWORKS. ACTA ACUST UNITED AC 2011. [DOI: 10.1142/s179300570600049x] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Unlike traditional artificial neural networks (ANNs), which use a high abstraction of real neurons, spiking neural networks (SNNs) offer a biologically plausible model of realistic neurons. They differ from classical artificial neural networks in that SNNs handle and communicate information by means of timing of individual pulses, an important feature of neuronal systems being ignored by models based on rate coding scheme. However, in order to make the most of these realistic neuronal models, good training algorithms are required. Most existing learning paradigms tune the synaptic weights in an unsupervised way using an adaptation of the famous Hebbian learning rule, which is based on the correlation between the pre- and post-synaptic neurons activity. Nonetheless, supervised learning is more appropriate when prior knowledge about the outcome of the network is available. In this paper, a new approach for supervised training is presented with a biologically plausible architecture. An adapted evolutionary strategy (ES) is used for adjusting the synaptic strengths and delays, which underlie the learning and memory processes in the nervous system. The algorithm is applied to complex non-linearly separable problems, and the results show that the network is able to perform learning successfully by means of temporal encoding of presented patterns.
Collapse
Affiliation(s)
- AMMAR BELATRECHE
- Intelligent Systems Engineering Laboratory, School of Computing and Intelligent Systems, University of Ulster, Magee Campus, Northland Road, Derry, BT48 7JL, Northern Ireland
| | - LIAM P. MAGUIRE
- Intelligent Systems Engineering Laboratory, School of Computing and Intelligent Systems, University of Ulster, Magee Campus, Northland Road, Derry, BT48 7JL, Northern Ireland
| | - MARTIN MCGINNITY
- Intelligent Systems Engineering Laboratory, School of Computing and Intelligent Systems, University of Ulster, Magee Campus, Northland Road, Derry, BT48 7JL, Northern Ireland
| | - QING XIANG WU
- Intelligent Systems Engineering Laboratory, School of Computing and Intelligent Systems, University of Ulster, Magee Campus, Northland Road, Derry, BT48 7JL, Northern Ireland
| |
Collapse
|
39
|
Byrnes S, Burkitt AN, Grayden DB, Meffin H. Learning a Sparse Code for Temporal Sequences Using STDP and Sequence Compression. Neural Comput 2011; 23:2567-98. [DOI: 10.1162/neco_a_00184] [Citation(s) in RCA: 19] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/04/2022]
Abstract
A spiking neural network that learns temporal sequences is described. A sparse code in which individual neurons represent sequences and subsequences enables multiple sequences to be stored without interference. The network is founded on a model of sequence compression in the hippocampus that is robust to variation in sequence element duration and well suited to learn sequences through spike-timing dependent plasticity (STDP). Three additions to the sequence compression model underlie the sparse representation: synapses connecting the neurons of the network that are subject to STDP, a competitive plasticity rule so that neurons specialize to individual sequences, and neural depolarization after spiking so that neurons have a memory. The response to new sequence elements is determined by the neurons that have responded to the previous subsequence, according to the competitively learned synaptic connections. Numerical simulations show that the model can learn sets of intersecting sequences, presented with widely differing frequencies, with elements of varying duration.
Collapse
Affiliation(s)
- Sean Byrnes
- Bionic Ear Institute, East Melbourne, Victoria 3002, Australia, and Department of Electrical and Electronic Engineering, University of Melbourne, Victoria 3010, Australia
| | - Anthony N. Burkitt
- Department of Electrical and Electronic Engineering, University of Melbourne, Victoria 3010, Australia, and Bionic Ear Institute, East Melbourne, Victoria 3002, Australia
| | - David B. Grayden
- Department of Electrical and Electronic Engineering, University of Melbourne, Victoria 3010, Australia, and Bionic Ear Institute, East Melbourne, Victoria 3002, Australia
| | - Hamish Meffin
- NICTA and Department of Electrical and Electronic Engineering, University of Melbourne, Victoria 3010, Australia
| |
Collapse
|
40
|
Abstract
A central criticism of standard theoretical approaches to constructing stable, recurrent model networks is that the synaptic connection weights need to be finely-tuned. This criticism is severe because proposed rules for learning these weights have been shown to have various limitations to their biological plausibility. Hence it is unlikely that such rules are used to continuously fine-tune the network in vivo. We describe a learning rule that is able to tune synaptic weights in a biologically plausible manner. We demonstrate and test this rule in the context of the oculomotor integrator, showing that only known neural signals are needed to tune the weights. We demonstrate that the rule appropriately accounts for a wide variety of experimental results, and is robust under several kinds of perturbation. Furthermore, we show that the rule is able to achieve stability as good as or better than that provided by the linearly optimal weights often used in recurrent models of the integrator. Finally, we discuss how this rule can be generalized to tune a wide variety of recurrent attractor networks, such as those found in head direction and path integration systems, suggesting that it may be used to tune a wide variety of stable neural systems.
Collapse
|
41
|
An imperfect dopaminergic error signal can drive temporal-difference learning. PLoS Comput Biol 2011; 7:e1001133. [PMID: 21589888 PMCID: PMC3093351 DOI: 10.1371/journal.pcbi.1001133] [Citation(s) in RCA: 43] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2010] [Accepted: 04/06/2011] [Indexed: 12/03/2022] Open
Abstract
An open problem in the field of computational neuroscience is how to link synaptic plasticity to system-level learning. A promising framework in this context is temporal-difference (TD) learning. Experimental evidence that supports the hypothesis that the mammalian brain performs temporal-difference learning includes the resemblance of the phasic activity of the midbrain dopaminergic neurons to the TD error and the discovery that cortico-striatal synaptic plasticity is modulated by dopamine. However, as the phasic dopaminergic signal does not reproduce all the properties of the theoretical TD error, it is unclear whether it is capable of driving behavior adaptation in complex tasks. Here, we present a spiking temporal-difference learning model based on the actor-critic architecture. The model dynamically generates a dopaminergic signal with realistic firing rates and exploits this signal to modulate the plasticity of synapses as a third factor. The predictions of our proposed plasticity dynamics are in good agreement with experimental results with respect to dopamine, pre- and post-synaptic activity. An analytical mapping from the parameters of our proposed plasticity dynamics to those of the classical discrete-time TD algorithm reveals that the biological constraints of the dopaminergic signal entail a modified TD algorithm with self-adapting learning parameters and an adapting offset. We show that the neuronal network is able to learn a task with sparse positive rewards as fast as the corresponding classical discrete-time TD algorithm. However, the performance of the neuronal network is impaired with respect to the traditional algorithm on a task with both positive and negative rewards and breaks down entirely on a task with purely negative rewards. Our model demonstrates that the asymmetry of a realistic dopaminergic signal enables TD learning when learning is driven by positive rewards but not when driven by negative rewards. What are the physiological changes that take place in the brain when we solve a problem or learn a new skill? It is commonly assumed that behavior adaptations are realized on the microscopic level by changes in synaptic efficacies. However, this is hard to verify experimentally due to the difficulties of identifying the relevant synapses and monitoring them over long periods during a behavioral task. To address this question computationally, we develop a spiking neuronal network model of actor-critic temporal-difference learning, a variant of reinforcement learning for which neural correlates have already been partially established. The network learns a complex task by means of an internally generated reward signal constrained by recent findings on the dopaminergic system. Our model combines top-down and bottom-up modelling approaches to bridge the gap between synaptic plasticity and system-level learning. It paves the way for further investigations of the dopaminergic system in reward learning in the healthy brain and in pathological conditions such as Parkinson's disease, and can be used as a module in functional models based on brain-scale circuitry.
Collapse
|
42
|
Zamarreño-Ramos C, Camuñas-Mesa LA, Pérez-Carrasco JA, Masquelier T, Serrano-Gotarredona T, Linares-Barranco B. On spike-timing-dependent-plasticity, memristive devices, and building a self-learning visual cortex. Front Neurosci 2011; 5:26. [PMID: 21442012 PMCID: PMC3062969 DOI: 10.3389/fnins.2011.00026] [Citation(s) in RCA: 118] [Impact Index Per Article: 9.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2010] [Accepted: 02/19/2011] [Indexed: 11/13/2022] Open
Abstract
In this paper we present a very exciting overlap between emergent nanotechnology and neuroscience, which has been discovered by neuromorphic engineers. Specifically, we are linking one type of memristor nanotechnology devices to the biological synaptic update rule known as spike-time-dependent-plasticity (STDP) found in real biological synapses. Understanding this link allows neuromorphic engineers to develop circuit architectures that use this type of memristors to artificially emulate parts of the visual cortex. We focus on the type of memristors referred to as voltage or flux driven memristors and focus our discussions on a behavioral macro-model for such devices. The implementations result in fully asynchronous architectures with neurons sending their action potentials not only forward but also backward. One critical aspect is to use neurons that generate spikes of specific shapes. We will see how by changing the shapes of the neuron action potential spikes we can tune and manipulate the STDP learning rules for both excitatory and inhibitory synapses. We will see how neurons and memristors can be interconnected to achieve large scale spiking learning systems, that follow a type of multiplicative STDP learning rule. We will briefly extend the architectures to use three-terminal transistors with similar memristive behavior. We will illustrate how a V1 visual cortex layer can assembled and how it is capable of learning to extract orientations from visual data coming from a real artificial CMOS spiking retina observing real life scenes. Finally, we will discuss limitations of currently available memristors. The results presented are based on behavioral simulations and do not take into account non-idealities of devices and interconnects. The aim of this paper is to present, in a tutorial manner, an initial framework for the possible development of fully asynchronous STDP learning neuromorphic architectures exploiting two or three-terminal memristive type devices. All files used for the simulations are made available through the journal web site.
Collapse
Affiliation(s)
- Carlos Zamarreño-Ramos
- Mixed Signal Design, Instituto de Microelectrónica de Sevilla (IMSE–CNM–CSIC)Sevilla, Spain
| | - Luis A. Camuñas-Mesa
- Mixed Signal Design, Instituto de Microelectrónica de Sevilla (IMSE–CNM–CSIC)Sevilla, Spain
| | - Jose A. Pérez-Carrasco
- Mixed Signal Design, Instituto de Microelectrónica de Sevilla (IMSE–CNM–CSIC)Sevilla, Spain
| | | | | | | |
Collapse
|
43
|
Abstract
Neuronal circuitry is often considered a clean slate that can be dynamically and arbitrarily molded by experience. However, when we investigated synaptic connectivity in groups of pyramidal neurons in the neocortex, we found that both connectivity and synaptic weights were surprisingly predictable. Synaptic weights follow very closely the number of connections in a group of neurons, saturating after only 20% of possible connections are formed between neurons in a group. When we examined the network topology of connectivity between neurons, we found that the neurons cluster into small world networks that are not scale-free, with less than 2 degrees of separation. We found a simple clustering rule where connectivity is directly proportional to the number of common neighbors, which accounts for these small world networks and accurately predicts the connection probability between any two neurons. This pyramidal neuron network clusters into multiple groups of a few dozen neurons each. The neurons composing each group are surprisingly distributed, typically more than 100 μm apart, allowing for multiple groups to be interlaced in the same space. In summary, we discovered a synaptic organizing principle that groups neurons in a manner that is common across animals and hence, independent of individual experiences. We speculate that these elementary neuronal groups are prescribed Lego-like building blocks of perception and that acquired memory relies more on combining these elementary assemblies into higher-order constructs.
Collapse
|
44
|
Froemke RC, Letzkus JJ, Kampa BM, Hang GB, Stuart GJ. Dendritic synapse location and neocortical spike-timing-dependent plasticity. Front Synaptic Neurosci 2010; 2:29. [PMID: 21423515 PMCID: PMC3059711 DOI: 10.3389/fnsyn.2010.00029] [Citation(s) in RCA: 46] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2010] [Accepted: 06/27/2010] [Indexed: 11/30/2022] Open
Abstract
While it has been appreciated for decades that synapse location in the dendritic tree has a powerful influence on signal processing in neurons, the role of dendritic synapse location on the induction of long-term synaptic plasticity has only recently been explored. Here, we review recent work revealing how learning rules for spike-timing-dependent plasticity (STDP) in cortical neurons vary with the spatial location of synaptic input. A common principle appears to be that proximal synapses show conventional STDP, whereas distal inputs undergo plasticity according to novel learning rules. One crucial factor determining location-dependent STDP is the backpropagating action potential, which tends to decrease in amplitude and increase in width as it propagates into the dendritic tree of cortical neurons. We discuss additional location-dependent mechanisms as well as the functional implications of heterogeneous learning rules at different dendritic locations for the organization of synaptic inputs.
Collapse
Affiliation(s)
- Robert C Froemke
- Departments of Otolaryngology and Physiology/Neuroscience, Molecular Neurobiology Program, The Helen and Martin Kimmel Center for Biology and Medicine, Skirball Institute of Biomolecular Medicine, New York University School of Medicine New York, NY, USA
| | | | | | | | | |
Collapse
|
45
|
Shouval HZ, Wang SSH, Wittenberg GM. Spike timing dependent plasticity: a consequence of more fundamental learning rules. Front Comput Neurosci 2010; 4. [PMID: 20725599 PMCID: PMC2922937 DOI: 10.3389/fncom.2010.00019] [Citation(s) in RCA: 62] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2010] [Accepted: 06/07/2010] [Indexed: 11/13/2022] Open
Abstract
Spike timing dependent plasticity (STDP) is a phenomenon in which the precise timing of spikes affects the sign and magnitude of changes in synaptic strength. STDP is often interpreted as the comprehensive learning rule for a synapse - the "first law" of synaptic plasticity. This interpretation is made explicit in theoretical models in which the total plasticity produced by complex spike patterns results from a superposition of the effects of all spike pairs. Although such models are appealing for their simplicity, they can fail dramatically. For example, the measured single-spike learning rule between hippocampal CA3 and CA1 pyramidal neurons does not predict the existence of long-term potentiation one of the best-known forms of synaptic plasticity. Layers of complexity have been added to the basic STDP model to repair predictive failures, but they have been outstripped by experimental data. We propose an alternate first law: neural activity triggers changes in key biochemical intermediates, which act as a more direct trigger of plasticity mechanisms. One particularly successful model uses intracellular calcium as the intermediate and can account for many observed properties of bidirectional plasticity. In this formulation, STDP is not itself the basis for explaining other forms of plasticity, but is instead a consequence of changes in the biochemical intermediate, calcium. Eventually a mechanism-based framework for learning rules should include other messengers, discrete change at individual synapses, spread of plasticity among neighboring synapses, and priming of hidden processes that change a synapse's susceptibility to future change. Mechanism-based models provide a rich framework for the computational representation of synaptic plasticity.
Collapse
Affiliation(s)
- Harel Z Shouval
- Department of Neurobiology and Anatomy, The University of Texas Medical School at Houston Houston, TX, USA
| | | | | |
Collapse
|
46
|
Froemke RC, Debanne D, Bi GQ. Temporal modulation of spike-timing-dependent plasticity. Front Synaptic Neurosci 2010; 2:19. [PMID: 21423505 PMCID: PMC3059714 DOI: 10.3389/fnsyn.2010.00019] [Citation(s) in RCA: 39] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2010] [Accepted: 05/27/2010] [Indexed: 11/13/2022] Open
Abstract
Spike-timing-dependent plasticity (STDP) has attracted considerable experimental and theoretical attention over the last decade. In the most basic formulation, STDP provides a fundamental unit – a spike pair – for quantifying the induction of long-term changes in synaptic strength. However, many factors, both pre- and postsynaptic, can affect synaptic transmission and integration, especially when multiple spikes are considered. Here we review the experimental evidence for multiple types of nonlinear temporal interactions in STDP, focusing on the contributions of individual spike pairs, overall spike rate, and precise spike timing for modification of cortical and hippocampal excitatory synapses. We discuss the underlying processes that determine the specific learning rules at different synapses, such as postsynaptic excitability and short-term depression. Finally, we describe the success of efforts toward building predictive, quantitative models of how complex and natural spike trains induce long-term synaptic modifications.
Collapse
Affiliation(s)
- Robert C Froemke
- Molecular Neurobiology Program, Departments of Otolaryngology and Physiology/Neuroscience, The Helen and Martin Kimmel Center for Biology and Medicine, Skirball Institute of Biomolecular Medicine, New York University School of Medicine New York, NY, USA
| | | | | |
Collapse
|
47
|
Synaptic information transfer in computer models of neocortical columns. J Comput Neurosci 2010; 30:69-84. [PMID: 20556639 DOI: 10.1007/s10827-010-0253-4] [Citation(s) in RCA: 41] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2009] [Revised: 05/24/2010] [Accepted: 05/31/2010] [Indexed: 10/19/2022]
Abstract
Understanding the direction and quantity of information flowing in neuronal networks is a fundamental problem in neuroscience. Brains and neuronal networks must at the same time store information about the world and react to information in the world. We sought to measure how the activity of the network alters information flow from inputs to output patterns. Using neocortical column neuronal network simulations, we demonstrated that networks with greater internal connectivity reduced input/output correlations from excitatory synapses and decreased negative correlations from inhibitory synapses, measured by Kendall's τ correlation. Both of these changes were associated with reduction in information flow, measured by normalized transfer entropy (nTE). Information handling by the network reflected the degree of internal connectivity. With no internal connectivity, the feedforward network transformed inputs through nonlinear summation and thresholding. With greater connectivity strength, the recurrent network translated activity and information due to contribution of activity from intrinsic network dynamics. This dynamic contribution amounts to added information drawn from that stored in the network. At still higher internal synaptic strength, the network corrupted the external information, producing a state where little external information came through. The association of increased information retrieved from the network with increased gamma power supports the notion of gamma oscillations playing a role in information processing.
Collapse
|
48
|
Bernasconi F, Grivel J, Murray MM, Spierer L. Plastic brain mechanisms for attaining auditory temporal order judgment proficiency. Neuroimage 2010; 50:1271-9. [DOI: 10.1016/j.neuroimage.2010.01.016] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2009] [Revised: 01/04/2010] [Accepted: 01/06/2010] [Indexed: 10/20/2022] Open
|
49
|
Synchronous neural activity and memory formation. Curr Opin Neurobiol 2010; 20:150-5. [PMID: 20303255 DOI: 10.1016/j.conb.2010.02.006] [Citation(s) in RCA: 139] [Impact Index Per Article: 9.9] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2009] [Accepted: 02/17/2010] [Indexed: 11/23/2022]
Abstract
Accumulating evidence suggests that the synchronization of neuronal activity plays an important role in memory formation. In particular, several recent studies have demonstrated that enhanced synchronous activity within and among medial temporal lobe structures is correlated with increased memory performance in humans and animals. Modulations in rhythmic synchronization in the gamma-frequency (30-100 Hz) and theta-frequency (4-8 Hz) bands have been related to memory performance, and interesting relationships have been described between these oscillations that suggest a mechanism for inter-areal coupling. Neuronal synchronization has also been linked to spike timing-dependent plasticity, a cellular mechanism thought to underlie learning and memory. The available evidence suggests that neuronal synchronization modulates memory performance as well as potential cellular mechanisms of memory storage.
Collapse
|
50
|
Perceptron learning rule derived from spike-frequency adaptation and spike-time-dependent plasticity. Proc Natl Acad Sci U S A 2010; 107:4722-7. [PMID: 20167805 DOI: 10.1073/pnas.0909394107] [Citation(s) in RCA: 20] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
It is widely believed that sensory and motor processing in the brain is based on simple computational primitives rooted in cellular and synaptic physiology. However, many gaps remain in our understanding of the connections between neural computations and biophysical properties of neurons. Here, we show that synaptic spike-time-dependent plasticity (STDP) combined with spike-frequency adaptation (SFA) in a single neuron together approximate the well-known perceptron learning rule. Our calculations and integrate-and-fire simulations reveal that delayed inputs to a neuron endowed with STDP and SFA precisely instruct neural responses to earlier arriving inputs. We demonstrate this mechanism on a developmental example of auditory map formation guided by visual inputs, as observed in the external nucleus of the inferior colliculus (ICX) of barn owls. The interplay of SFA and STDP in model ICX neurons precisely transfers the tuning curve from the visual modality onto the auditory modality, demonstrating a useful computation for multimodal and sensory-guided processing.
Collapse
|