1
|
Capone C, Paolucci PS. Towards biologically plausible model-based reinforcement learning in recurrent spiking networks by dreaming new experiences. Sci Rep 2024; 14:14656. [PMID: 38918553 PMCID: PMC11199658 DOI: 10.1038/s41598-024-65631-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2024] [Accepted: 06/21/2024] [Indexed: 06/27/2024] Open
Abstract
Humans and animals can learn new skills after practicing for a few hours, while current reinforcement learning algorithms require a large amount of data to achieve good performances. Recent model-based approaches show promising results by reducing the number of necessary interactions with the environment to learn a desirable policy. However, these methods require biological implausible ingredients, such as the detailed storage of older experiences, and long periods of offline learning. The optimal way to learn and exploit world-models is still an open question. Taking inspiration from biology, we suggest that dreaming might be an efficient expedient to use an inner model. We propose a two-module (agent and model) spiking neural network in which "dreaming" (living new experiences in a model-based simulated environment) significantly boosts learning. Importantly, our model does not require the detailed storage of experiences, and learns online the world-model and the policy. Moreover, we stress that our network is composed of spiking neurons, further increasing the biological plausibility and implementability in neuromorphic hardware.
Collapse
Affiliation(s)
- Cristiano Capone
- INFN, Sezione di Roma, Rome, RM, 00185, Italy.
- Natl. Center for Radiation Protection and Computational Physics, Istituto Superiore di Sanitá, Rome, 00161, Italy.
| | | |
Collapse
|
2
|
Li F, Li D, Wang C, Liu G, Wang R, Ren H, Tang Y, Wang Y, Chen Y, Liang K, Huang Q, Sawan M, Qiu M, Wang H, Zhu B. An artificial visual neuron with multiplexed rate and time-to-first-spike coding. Nat Commun 2024; 15:3689. [PMID: 38693165 PMCID: PMC11063071 DOI: 10.1038/s41467-024-48103-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Accepted: 04/17/2024] [Indexed: 05/03/2024] Open
Abstract
Human visual neurons rely on event-driven, energy-efficient spikes for communication, while silicon image sensors do not. The energy-budget mismatch between biological systems and machine vision technology has inspired the development of artificial visual neurons for use in spiking neural network (SNN). However, the lack of multiplexed data coding schemes reduces the ability of artificial visual neurons in SNN to emulate the visual perception ability of biological systems. Here, we present an artificial visual spiking neuron that enables rate and temporal fusion (RTF) coding of external visual information. The artificial neuron can code visual information at different spiking frequencies (rate coding) and enables precise and energy-efficient time-to-first-spike (TTFS) coding. This multiplexed sensory coding scheme could improve the computing capability and efficacy of artificial visual neurons. A hardware-based SNN with the RTF coding scheme exhibits good consistency with real-world ground truth data and achieves highly accurate steering and speed predictions for self-driving vehicles in complex conditions. The multiplexed RTF coding scheme demonstrates the feasibility of developing highly efficient spike-based neuromorphic hardware.
Collapse
Affiliation(s)
- Fanfan Li
- School of Materials Science and Engineering, Zhejiang University, Hangzhou, China
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Dingwei Li
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Chuanqing Wang
- CenBRAIN Neurotech, School of Engineering, Westlake University, Hangzhou, China
| | - Guolei Liu
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Rui Wang
- Key Laboratory of Wide Band Gap Semiconductor Technology, School of Microelectronics, Xidian University, Xi'an, China
| | - Huihui Ren
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Yingjie Tang
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Yan Wang
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Yitong Chen
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Kun Liang
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
| | - Qi Huang
- Westlake Institute for Optoelectronics, Westlake University, Hangzhou, China
| | - Mohamad Sawan
- CenBRAIN Neurotech, School of Engineering, Westlake University, Hangzhou, China
- Westlake Institute for Optoelectronics, Westlake University, Hangzhou, China
- Institute of Advanced Technology, Westlake Institute for Advanced Study, Hangzhou, China
| | - Min Qiu
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China
- Westlake Institute for Optoelectronics, Westlake University, Hangzhou, China
- Institute of Advanced Technology, Westlake Institute for Advanced Study, Hangzhou, China
| | - Hong Wang
- Key Laboratory of Wide Band Gap Semiconductor Technology, School of Microelectronics, Xidian University, Xi'an, China.
| | - Bowen Zhu
- Key Laboratory of 3D Micro/Nano Fabrication and Characterization of Zhejiang Province, School of Engineering, Westlake University, Hangzhou, China.
- Westlake Institute for Optoelectronics, Westlake University, Hangzhou, China.
- Institute of Advanced Technology, Westlake Institute for Advanced Study, Hangzhou, China.
| |
Collapse
|
3
|
Sakemi Y, Yamamoto K, Hosomi T, Aihara K. Sparse-firing regularization methods for spiking neural networks with time-to-first-spike coding. Sci Rep 2023; 13:22897. [PMID: 38129555 PMCID: PMC10739753 DOI: 10.1038/s41598-023-50201-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 12/16/2023] [Indexed: 12/23/2023] Open
Abstract
The training of multilayer spiking neural networks (SNNs) using the error backpropagation algorithm has made significant progress in recent years. Among the various training schemes, the error backpropagation method that directly uses the firing time of neurons has attracted considerable attention because it can realize ideal temporal coding. This method uses time-to-first-spike (TTFS) coding, in which each neuron fires at most once, and this restriction on the number of firings enables information to be processed at a very low firing frequency. This low firing frequency increases the energy efficiency of information processing in SNNs. However, only an upper limit has been provided for TTFS-coded SNNs, and the information-processing capability of SNNs at lower firing frequencies has not been fully investigated. In this paper, we propose two spike-timing-based sparse-firing (SSR) regularization methods to further reduce the firing frequency of TTFS-coded SNNs. Both methods are characterized by the fact that they only require information about the firing timing and associated weights. The effects of these regularization methods were investigated on the MNIST, Fashion-MNIST, and CIFAR-10 datasets using multilayer perceptron networks and convolutional neural network structures.
Collapse
Affiliation(s)
- Yusuke Sakemi
- Research Center for Mathematical Engineering, Chiba Institute of Technology, Narashino, Japan.
- International Research Center for Neurointelligence (WPI-IRCN), The University of Tokyo, Tokyo, Japan.
| | | | | | - Kazuyuki Aihara
- Research Center for Mathematical Engineering, Chiba Institute of Technology, Narashino, Japan
- International Research Center for Neurointelligence (WPI-IRCN), The University of Tokyo, Tokyo, Japan
| |
Collapse
|
4
|
Stanojevic A, Woźniak S, Bellec G, Cherubini G, Pantazi A, Gerstner W. An exact mapping from ReLU networks to spiking neural networks. Neural Netw 2023; 168:74-88. [PMID: 37742533 DOI: 10.1016/j.neunet.2023.09.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 08/31/2023] [Accepted: 09/04/2023] [Indexed: 09/26/2023]
Abstract
Deep spiking neural networks (SNNs) offer the promise of low-power artificial intelligence. However, training deep SNNs from scratch or converting deep artificial neural networks to SNNs without loss of performance has been a challenge. Here we propose an exact mapping from a network with Rectified Linear Units (ReLUs) to an SNN that fires exactly one spike per neuron. For our constructive proof, we assume that an arbitrary multi-layer ReLU network with or without convolutional layers, batch normalization and max pooling layers was trained to high performance on some training set. Furthermore, we assume that we have access to a representative example of input data used during training and to the exact parameters (weights and biases) of the trained ReLU network. The mapping from deep ReLU networks to SNNs causes zero percent drop in accuracy on CIFAR10, CIFAR100 and the ImageNet-like data sets Places365 and PASS. More generally our work shows that an arbitrary deep ReLU network can be replaced by an energy-efficient single-spike neural network without any loss of performance.
Collapse
Affiliation(s)
- Ana Stanojevic
- IBM Research Europe - Zurich, Rüschlikon, Switzerland; École polytechnique fédérale de Lausanne, School of Life Sciences and School of Computer and Communication Sciences, Lausanne EPFL, Switzerland.
| | | | - Guillaume Bellec
- École polytechnique fédérale de Lausanne, School of Life Sciences and School of Computer and Communication Sciences, Lausanne EPFL, Switzerland
| | | | | | - Wulfram Gerstner
- École polytechnique fédérale de Lausanne, School of Life Sciences and School of Computer and Communication Sciences, Lausanne EPFL, Switzerland
| |
Collapse
|
5
|
Liu S, Leung VCH, Dragotti PL. First-spike coding promotes accurate and efficient spiking neural networks for discrete events with rich temporal structures. Front Neurosci 2023; 17:1266003. [PMID: 37849889 PMCID: PMC10577212 DOI: 10.3389/fnins.2023.1266003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Accepted: 09/11/2023] [Indexed: 10/19/2023] Open
Abstract
Spiking neural networks (SNNs) are well-suited to process asynchronous event-based data. Most of the existing SNNs use rate-coding schemes that focus on firing rate (FR), and so they generally ignore the spike timing in events. On the contrary, methods based on temporal coding, particularly time-to-first-spike (TTFS) coding, can be accurate and efficient but they are difficult to train. Currently, there is limited research on applying TTFS coding to real events, since traditional TTFS-based methods impose one-spike constraint, which is not realistic for event-based data. In this study, we present a novel decision-making strategy based on first-spike (FS) coding that encodes FS timings of the output neurons to investigate the role of the first-spike timing in classifying real-world event sequences with complex temporal structures. To achieve FS coding, we propose a novel surrogate gradient learning method for discrete spike trains. In the forward pass, output spikes are encoded into discrete times to generate FS times. In the backpropagation, we develop an error assignment method that propagates error from FS times to spikes through a Gaussian window, and then supervised learning for spikes is implemented through a surrogate gradient approach. Additional strategies are introduced to facilitate the training of FS timings, such as adding empty sequences and employing different parameters for different layers. We make a comprehensive comparison between FS and FR coding in the experiments. Our results show that FS coding achieves comparable accuracy to FR coding while leading to superior energy efficiency and distinct neuronal dynamics on data sequences with very rich temporal structures. Additionally, a longer time delay in the first spike leads to higher accuracy, indicating important information is encoded in the timing of the first spike.
Collapse
Affiliation(s)
- Siying Liu
- Communications and Signal Processing Group, Department of Electrical and Electronic Engineering, Imperial College London, London, United Kingdom
| | | | | |
Collapse
|
6
|
Bacho F, Chu D. Exploring Trade-Offs in Spiking Neural Networks. Neural Comput 2023; 35:1627-1656. [PMID: 37523463 DOI: 10.1162/neco_a_01609] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Accepted: 06/03/2023] [Indexed: 08/02/2023]
Abstract
Spiking neural networks (SNNs) have emerged as a promising alternative to traditional deep neural networks for low-power computing. However, the effectiveness of SNNs is not solely determined by their performance but also by their energy consumption, prediction speed, and robustness to noise. The recent method Fast & Deep, along with others, achieves fast and energy-efficient computation by constraining neurons to fire at most once. Known as time-to-first-spike (TTFS), this constraint, however, restricts the capabilities of SNNs in many aspects. In this work, we explore the relationships of performance, energy consumption, speed, and stability when using this constraint. More precisely, we highlight the existence of trade-offs where performance and robustness are gained at the cost of sparsity and prediction latency. To improve these trade-offs, we propose a relaxed version of Fast & Deep that allows for multiple spikes per neuron. Our experiments show that relaxing the spike constraint provides higher performance while also benefiting from faster convergence, similar sparsity, comparable prediction latency, and better robustness to noise compared to TTFS SNNs. By highlighting the limitations of TTFS and demonstrating the advantages of unconstrained SNNs, we provide valuable insight for the development of effective learning strategies for neuromorphic computing.
Collapse
Affiliation(s)
- Florian Bacho
- CEMS, School of Computing, University of Kent, Canterbury CT2 7NF, U.K.
| | - Dominique Chu
- CEMS, School of Computing, University of Kent, Canterbury CT2 7NF, U.K.
| |
Collapse
|
7
|
Sanders LM, Scott RT, Yang JH, Qutub AA, Garcia Martin H, Berrios DC, Hastings JJA, Rask J, Mackintosh G, Hoarfrost AL, Chalk S, Kalantari J, Khezeli K, Antonsen EL, Babdor J, Barker R, Baranzini SE, Beheshti A, Delgado-Aparicio GM, Glicksberg BS, Greene CS, Haendel M, Hamid AA, Heller P, Jamieson D, Jarvis KJ, Komarova SV, Komorowski M, Kothiyal P, Mahabal A, Manor U, Mason CE, Matar M, Mias GI, Miller J, Myers JG, Nelson C, Oribello J, Park SM, Parsons-Wingerter P, Prabhu RK, Reynolds RJ, Saravia-Butler A, Saria S, Sawyer A, Singh NK, Snyder M, Soboczenski F, Soman K, Theriot CA, Van Valen D, Venkateswaran K, Warren L, Worthey L, Zitnik M, Costes SV. Biological research and self-driving labs in deep space supported by artificial intelligence. NAT MACH INTELL 2023. [DOI: 10.1038/s42256-023-00618-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/28/2023]
|
8
|
Bauer FC, Lenz G, Haghighatshoar S, Sheik S. EXODUS: Stable and efficient training of spiking neural networks. Front Neurosci 2023; 17:1110444. [PMID: 36845419 PMCID: PMC9945199 DOI: 10.3389/fnins.2023.1110444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Accepted: 01/09/2023] [Indexed: 02/10/2023] Open
Abstract
Introduction Spiking Neural Networks (SNNs) are gaining significant traction in machine learning tasks where energy-efficiency is of utmost importance. Training such networks using the state-of-the-art back-propagation through time (BPTT) is, however, very time-consuming. Previous work employs an efficient GPU-accelerated backpropagation algorithm called SLAYER, which speeds up training considerably. SLAYER, however, does not take into account the neuron reset mechanism while computing the gradients, which we argue to be the source of numerical instability. To counteract this, SLAYER introduces a gradient scale hyper parameter across layers, which needs manual tuning. Methods In this paper, we modify SLAYER and design an algorithm called EXODUS, that accounts for the neuron reset mechanism and applies the Implicit Function Theorem (IFT) to calculate the correct gradients (equivalent to those computed by BPTT). We furthermore eliminate the need for ad-hoc scaling of gradients, thus, reducing the training complexity tremendously. Results We demonstrate, via computer simulations, that EXODUS is numerically stable and achieves comparable or better performance than SLAYER especially in various tasks with SNNs that rely on temporal features.
Collapse
|
9
|
Pyo J, Bae JH, Kim S, Cho S. Short-Term Memory Characteristics of IGZO-Based Three-Terminal Devices. MATERIALS (BASEL, SWITZERLAND) 2023; 16:1249. [PMID: 36770256 PMCID: PMC9919079 DOI: 10.3390/ma16031249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Revised: 01/19/2023] [Accepted: 01/26/2023] [Indexed: 06/18/2023]
Abstract
A three-terminal synaptic transistor enables more accurate controllability over the conductance compared with traditional two-terminal synaptic devices for the synaptic devices in hardware-oriented neuromorphic systems. In this work, we fabricated IGZO-based three-terminal devices comprising HfAlOx and CeOx layers to demonstrate the synaptic operations. The chemical compositions and thicknesses of the devices were verified by transmission electron microscopy and energy dispersive spectroscopy in cooperation. The excitatory post-synaptic current (EPSC), paired-pulse facilitation (PPF), short-term potentiation (STP), and short-term depression (STD) of the synaptic devices were realized for the short-term memory behaviors. The IGZO-based three-terminal synaptic transistor could thus be controlled appropriately by the amplitude, width, and interval time of the pulses for implementing the neuromorphic systems.
Collapse
Affiliation(s)
- Juyeong Pyo
- Division of Electronics and Electrical Engineering, Dongguk University, Seoul 04620, Republic of Korea
| | - Jong-Ho Bae
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Sungjun Kim
- Division of Electronics and Electrical Engineering, Dongguk University, Seoul 04620, Republic of Korea
| | - Seongjae Cho
- Department of Electronics Engineering, Gachon University, Seongnam 13120, Republic of Korea
| |
Collapse
|
10
|
Konar D, Sarma AD, Bhandary S, Bhattacharyya S, Cangi A, Aggarwal V. A shallow hybrid classical-quantum spiking feedforward neural network for noise-robust image classification. Appl Soft Comput 2023. [DOI: 10.1016/j.asoc.2023.110099] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/12/2023]
|
11
|
Nilsson M, Schelén O, Lindgren A, Bodin U, Paniagua C, Delsing J, Sandin F. Integration of neuromorphic AI in event-driven distributed digitized systems: Concepts and research directions. Front Neurosci 2023; 17:1074439. [PMID: 36875653 PMCID: PMC9981939 DOI: 10.3389/fnins.2023.1074439] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 01/23/2023] [Indexed: 02/19/2023] Open
Abstract
Increasing complexity and data-generation rates in cyber-physical systems and the industrial Internet of things are calling for a corresponding increase in AI capabilities at the resource-constrained edges of the Internet. Meanwhile, the resource requirements of digital computing and deep learning are growing exponentially, in an unsustainable manner. One possible way to bridge this gap is the adoption of resource-efficient brain-inspired "neuromorphic" processing and sensing devices, which use event-driven, asynchronous, dynamic neurosynaptic elements with colocated memory for distributed processing and machine learning. However, since neuromorphic systems are fundamentally different from conventional von Neumann computers and clock-driven sensor systems, several challenges are posed to large-scale adoption and integration of neuromorphic devices into the existing distributed digital-computational infrastructure. Here, we describe the current landscape of neuromorphic computing, focusing on characteristics that pose integration challenges. Based on this analysis, we propose a microservice-based conceptual framework for neuromorphic systems integration, consisting of a neuromorphic-system proxy, which would provide virtualization and communication capabilities required in distributed systems of systems, in combination with a declarative programming approach offering engineering-process abstraction. We also present concepts that could serve as a basis for the realization of this framework, and identify directions for further research required to enable large-scale system integration of neuromorphic devices.
Collapse
Affiliation(s)
- Mattias Nilsson
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| | - Olov Schelén
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| | - Anders Lindgren
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden.,Applied AI and IoT, Industrial Systems, Digital Systems, RISE Research Institutes of Sweden, Kista, Sweden
| | - Ulf Bodin
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| | - Cristina Paniagua
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| | - Jerker Delsing
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| | - Fredrik Sandin
- Embedded Intelligent Systems Lab (EISLAB), Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, Lulea, Sweden
| |
Collapse
|
12
|
Dendrocentric learning for synthetic intelligence. Nature 2022; 612:43-50. [DOI: 10.1038/s41586-022-05340-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Accepted: 09/12/2022] [Indexed: 12/02/2022]
|
13
|
Redolfi A, Archetti D, De Francesco S, Crema C, Tagliavini F, Lodi R, Ghidoni R, Gandini Wheeler-Kingshott CAM, Alexander DC, D'Angelo E. Italian, European, and international neuroinformatics efforts: An overview. Eur J Neurosci 2022. [PMID: 36310103 DOI: 10.1111/ejn.15854] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2022] [Revised: 10/18/2022] [Accepted: 10/18/2022] [Indexed: 12/15/2022]
Abstract
Neuroinformatics is a research field that focusses on software tools capable of identifying, analysing, modelling, organising and sharing multiscale neuroscience data. Neuroinformatics has exploded in the last two decades with the emergence of the Big Data phenomenon, characterised by the so-called 3Vs (volume, velocity and variety), which provided neuroscientists with an improved ability to acquire and process data faster and more cheaply thanks to technical improvements in clinical, genomic and radiological technologies. This situation has led to a 'data deluge', as neuroscientists can routinely collect more study data in a few days than they could in a year just a decade ago. To address this phenomenon, several neuroimaging-focussed neuroinformatics platforms have emerged, funded by national or transnational agencies, with the following goals: (i) development of tools for archiving and organising analytical data (XNAT, REDCap and LabKey); (ii) development of data-driven models evolving from reductionist approaches to multidimensional models (RIN, IVN, HBD, EuroPOND, E-DADS and GAAIN BRAIN); and (iii) development of e-infrastructures to provide sufficient computational power and storage resources (neuGRID, HBP-EBRAINS, LONI and CONP). Although the scenario is still fragmented, there are technological and economical attempts at both national and international levels to introduce high standards for open and Findable, Accessible, Interoperable and Reusable (FAIR) neuroscience worldwide.
Collapse
Affiliation(s)
- Alberto Redolfi
- Laboratory of Neuroinformatics, IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli, Brescia, Italy
| | - Damiano Archetti
- Laboratory of Neuroinformatics, IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli, Brescia, Italy
| | - Silvia De Francesco
- Laboratory of Neuroinformatics, IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli, Brescia, Italy
| | - Claudio Crema
- Laboratory of Neuroinformatics, IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli, Brescia, Italy
| | - Fabrizio Tagliavini
- Scientific Directorate, Fondazione IRCCS Istituto Neurologico Carlo Besta, Milan, Italy
| | - Raffaele Lodi
- Functional and Molecular Neuroimaging Unit, IRCCS Istituto delle Scienze Neurologiche di Bologna, Bologna, Italy.,Department of Biomedical and Neuromotor Sciences, University of Bologna, Bologna, Italy
| | - Roberta Ghidoni
- Molecular Markers Laboratory, IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli, Brescia, Italy
| | - Claudia A M Gandini Wheeler-Kingshott
- NMR Research Unit, Queen Square MS Center, Department of Neuroinflammation, UCL Institute of Neurology, London, UK.,Department of Brain and Behavioral Sciences, University of Pavia, Pavia, Italy.,Brain Connectivity Center, IRCCS Mondino Foundation, Pavia, Italy
| | - Daniel C Alexander
- Centre for Medical Image Computing, University College London, London, UK.,Department of Computer Science, University College London, London, UK
| | - Egidio D'Angelo
- Department of Brain and Behavioral Sciences, University of Pavia, Pavia, Italy.,Brain Connectivity Center, IRCCS Mondino Foundation, Pavia, Italy
| |
Collapse
|
14
|
Bonilla L, Gautrais J, Thorpe S, Masquelier T. Analyzing time-to-first-spike coding schemes: A theoretical approach. Front Neurosci 2022; 16:971937. [PMID: 36225737 PMCID: PMC9548614 DOI: 10.3389/fnins.2022.971937] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2022] [Accepted: 08/26/2022] [Indexed: 11/14/2022] Open
Abstract
Spiking neural networks (SNNs) using time-to-first-spike (TTFS) codes, in which neurons fire at most once, are appealing for rapid and low power processing. In this theoretical paper, we focus on information coding and decoding in those networks, and introduce a new unifying mathematical framework that allows the comparison of various coding schemes. In an early proposal, called rank-order coding (ROC), neurons are maximally activated when inputs arrive in the order of their synaptic weights, thanks to a shunting inhibition mechanism that progressively desensitizes the neurons as spikes arrive. In another proposal, called NoM coding, only the first N spikes of M input neurons are propagated, and these “first spike patterns” can be readout by downstream neurons with homogeneous weights and no desensitization: as a result, the exact order between the first spikes does not matter. This paper also introduces a third option—“Ranked-NoM” (R-NoM), which combines features from both ROC and NoM coding schemes: only the first N input spikes are propagated, but their order is readout by downstream neurons thanks to inhomogeneous weights and linear desensitization. The unifying mathematical framework allows the three codes to be compared in terms of discriminability, which measures to what extent a neuron responds more strongly to its preferred input spike pattern than to random patterns. This discriminability turns out to be much higher for R-NoM than for the other codes, especially in the early phase of the responses. We also argue that R-NoM is much more hardware-friendly than the original ROC proposal, although NoM remains the easiest to implement in hardware because it only requires binary synapses.
Collapse
Affiliation(s)
- Lina Bonilla
- CERCO UMR5549, CNRS – Université Toulouse III, Toulouse, France
- *Correspondence: Lina Bonilla
| | - Jacques Gautrais
- Centre de Recherches sur la Cognition Animale (CRCA), Centre de Biologie Intégrative (CBI), Université de Toulouse, Toulouse, France
- CNRS, UPS, Toulouse, France
| | - Simon Thorpe
- CERCO UMR5549, CNRS – Université Toulouse III, Toulouse, France
| | | |
Collapse
|
15
|
Müller E, Schmitt S, Mauch C, Billaudelle S, Grübl A, Güttler M, Husmann D, Ilmberger J, Jeltsch S, Kaiser J, Klähn J, Kleider M, Koke C, Montes J, Müller P, Partzsch J, Passenberg F, Schmidt H, Vogginger B, Weidner J, Mayr C, Schemmel J. The operating system of the neuromorphic BrainScaleS-1 system. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.05.081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
16
|
Yang H, Lam KY, Xiao L, Xiong Z, Hu H, Niyato D, Vincent Poor H. Lead federated neuromorphic learning for wireless edge artificial intelligence. Nat Commun 2022; 13:4269. [PMID: 35879326 PMCID: PMC9314401 DOI: 10.1038/s41467-022-32020-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2022] [Accepted: 07/13/2022] [Indexed: 12/02/2022] Open
Abstract
In order to realize the full potential of wireless edge artificial intelligence (AI), very large and diverse datasets will often be required for energy-demanding model training on resource-constrained edge devices. This paper proposes a lead federated neuromorphic learning (LFNL) technique, which is a decentralized energy-efficient brain-inspired computing method based on spiking neural networks. The proposed technique will enable edge devices to exploit brain-like biophysiological structure to collaboratively train a global model while helping preserve privacy. Experimental results show that, under the situation of uneven dataset distribution among edge devices, LFNL achieves a comparable recognition accuracy to existing edge AI techniques, while substantially reducing data traffic by >3.5× and computational latency by >2.0×. Furthermore, LFNL significantly reduces energy consumption by >4.5× compared to standard federated learning with a slight accuracy loss up to 1.5%. Therefore, the proposed LFNL can facilitate the development of brain-inspired computing and edge AI. Designing energy-efficient computing solution for the implementation of AI algorithms in edge devices remains a challenge. Yang et al. proposes a decentralized brain-inspired computing method enabling multiple edge devices to collaboratively train a global model without a fixed central coordinator.
Collapse
Affiliation(s)
- Helin Yang
- Department of Information and Communication Engineering, School of Informatics, Xiamen University, Xiamen, China.,Strategic Centre for Research in Privacy-Preserving Technologies and Systems, Nanyang Technological University, Singapore, Singapore
| | - Kwok-Yan Lam
- Strategic Centre for Research in Privacy-Preserving Technologies and Systems, Nanyang Technological University, Singapore, Singapore. .,School of Computer Science and Engineering, Nanyang Technological University, Singapore, Singapore.
| | - Liang Xiao
- Department of Information and Communication Engineering, School of Informatics, Xiamen University, Xiamen, China
| | - Zehui Xiong
- Pillar of Information Systems Technology and Design, Singapore University of Technology and Design, Singapore, Singapore
| | - Hao Hu
- Department of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Dusit Niyato
- School of Computer Science and Engineering, Nanyang Technological University, Singapore, Singapore
| | - H Vincent Poor
- Department of Electrical and Computer Engineering, Princeton University, Princeton, NJ, USA
| |
Collapse
|
17
|
Klassert R, Baumbach A, Petrovici MA, Gärttner M. Variational learning of quantum ground states on spiking neuromorphic hardware. iScience 2022; 25:104707. [PMID: 35992070 PMCID: PMC9386107 DOI: 10.1016/j.isci.2022.104707] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2022] [Revised: 05/05/2022] [Accepted: 06/28/2022] [Indexed: 11/26/2022] Open
Abstract
Recent research has demonstrated the usefulness of neural networks as variational ansatz functions for quantum many-body states. However, high-dimensional sampling spaces and transient autocorrelations confront these approaches with a challenging computational bottleneck. Compared to conventional neural networks, physical model devices offer a fast, efficient and inherently parallel substrate capable of related forms of Markov chain Monte Carlo sampling. Here, we demonstrate the ability of a neuromorphic chip to represent the ground states of quantum spin models by variational energy minimization. We develop a training algorithm and apply it to the transverse field Ising model, showing good performance at moderate system sizes (N≤10). A systematic hyperparameter study shows that performance depends on sample quality, which is limited by temporal parameter variations on the analog neuromorphic chip. Our work thus provides an important step towards harnessing the capabilities of neuromorphic hardware for tackling the curse of dimensionality in quantum many-body problems. Variational scheme for representing quantum ground states with neuromorphic hardware Accelerated physical system yields system-size independent sample generation time Accurate learning of ground states across a quantum phase transition Detailed analysis of algorithmic and technical limitations
Collapse
|
18
|
Wang X, Li H. A complementary resistive switching neuron. NANOTECHNOLOGY 2022; 33:355201. [PMID: 35605579 DOI: 10.1088/1361-6528/ac7241] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/22/2022] [Accepted: 05/22/2022] [Indexed: 06/15/2023]
Abstract
The complementary resistive switching (CRS) memristor has originally been proposed for use as the storage element or artificial synapse in large-scale crossbar array with the capability of solving the sneak path problem, but its usage has mainly been hampered by the inherent destructiveness of the read operation (switching '1' state to 'ON' or '0' state). Taking a different perspective on this 'undesired' property, we here report on the inherent behavioral similarity between the CRS memristor and a leaky integrate-and-fire (LIF) neuron which is another basic neural computing element, in addition to synapse. In particular, the mechanism behind the undesired read destructiveness for storage element and artificial synapse can be exploited to naturally realize the LIF and the ensuing spontaneous repolarization processes, followed by a refractory period. By means of this biological similarity, we demonstrate a Pt/Ta2O5-x/TaOy/Ta CRS memristor that can exhibit these neuronal behaviors and perform various fundamental neuronal operations, including additive/subtractive operations and coincidence detection. These results suggest that the CRS neuron, with its bio-interpretability, is a useful addition to the family of memristive neurons.
Collapse
Affiliation(s)
- Xinxin Wang
- Department of Precision Instrument, Center for Brain Inspired Computing Research, Tsinghua University, Beijing, 100084, People's Republic of China
| | - Huanglong Li
- Department of Precision Instrument, Center for Brain Inspired Computing Research, Tsinghua University, Beijing, 100084, People's Republic of China
- Chinese Institute for Brain Research, Beijing 102206, People's Republic of China
| |
Collapse
|
19
|
Müller E, Arnold E, Breitwieser O, Czierlinski M, Emmel A, Kaiser J, Mauch C, Schmitt S, Spilger P, Stock R, Stradmann Y, Weis J, Baumbach A, Billaudelle S, Cramer B, Ebert F, Göltz J, Ilmberger J, Karasenko V, Kleider M, Leibfried A, Pehle C, Schemmel J. A Scalable Approach to Modeling on Accelerated Neuromorphic Hardware. Front Neurosci 2022; 16:884128. [PMID: 35663548 PMCID: PMC9157770 DOI: 10.3389/fnins.2022.884128] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Accepted: 04/20/2022] [Indexed: 11/29/2022] Open
Abstract
Neuromorphic systems open up opportunities to enlarge the explorative space for computational research. However, it is often challenging to unite efficiency and usability. This work presents the software aspects of this endeavor for the BrainScaleS-2 system, a hybrid accelerated neuromorphic hardware architecture based on physical modeling. We introduce key aspects of the BrainScaleS-2 Operating System: experiment workflow, API layering, software design, and platform operation. We present use cases to discuss and derive requirements for the software and showcase the implementation. The focus lies on novel system and software features such as multi-compartmental neurons, fast re-configuration for hardware-in-the-loop training, applications for the embedded processors, the non-spiking operation mode, interactive platform access, and sustainable hardware/software co-development. Finally, we discuss further developments in terms of hardware scale-up, system usability, and efficiency.
Collapse
Affiliation(s)
- Eric Müller
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Elias Arnold
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Oliver Breitwieser
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Milena Czierlinski
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Arne Emmel
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Jakob Kaiser
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Christian Mauch
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Sebastian Schmitt
- Third Institute of Physics, University of Göttingen, Göttingen, Germany
| | - Philipp Spilger
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Raphael Stock
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Yannik Stradmann
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Johannes Weis
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Andreas Baumbach
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
- Department of Physiology, University of Bern, Bern, Switzerland
| | | | - Benjamin Cramer
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Falk Ebert
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Julian Göltz
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
- Department of Physiology, University of Bern, Bern, Switzerland
| | - Joscha Ilmberger
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Vitali Karasenko
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Mitja Kleider
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Aron Leibfried
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Christian Pehle
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| | - Johannes Schemmel
- Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| |
Collapse
|
20
|
Zhang M, Gao Q. Online Course Model of Social and Political Education Using Deep Learning. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:7653766. [PMID: 35498168 PMCID: PMC9042604 DOI: 10.1155/2022/7653766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/28/2021] [Revised: 01/11/2022] [Accepted: 01/17/2022] [Indexed: 11/17/2022]
Abstract
This study aims to improve the social and political literacy of college students. Social and Political Education (SPE) is studied for undergraduates. Firstly, the background of the subject research is introduced. The face recognition module is built based on deep convolutional neural network (DCNN). The sociopolitical situation of the study subjects is analyzed through a questionnaire. Secondly, a model of the learning process is constructed. Finally, the SPE online course learning platform is constructed. Empirical studies are divided into experimental and control groups. The findings show that all model assumptions are valid. There is a significant structural relationship between the influencing factors of the SPE learning process of college students in the study area. The students selected as research objects lack innovation and critical thinking in the learning process and have certain deficiencies in innovative thinking and critical thinking. The questionnaire has good reliability and validity. The predicted data of the designed platform are compared with the predicted data of the control group. Social science competencies by gender are compared. The results showed little difference in the effectiveness of students using other methods for sociopolitical learning. The data of the experimental group before and after the test are quite different, indicating that the designed experimental platform has played a certain positive role. There are significant differences in the posttest data between the experimental group and the control group, indicating that the constructed online course learning model has a positive impact on students' innovative thinking and critical thinking. Women's learning motivation and transfer learning ability are stronger than those of men. The constructed model has certain feasibility for the learning of SPE online courses with face recognition module. These contents provide a reference for the reform of social and political courses.
Collapse
Affiliation(s)
- Min Zhang
- Xi'an Technological University, Xi'an 710021, Shaanxi, China
| | - Qiong Gao
- Shandong University of Finance and Economics, Jinan 250014, Shandong, China
| |
Collapse
|
21
|
Vogginger B, Kreutz F, López-Randulfe J, Liu C, Dietrich R, Gonzalez HA, Scholz D, Reeb N, Auge D, Hille J, Arsalan M, Mirus F, Grassmann C, Knoll A, Mayr C. Automotive Radar Processing With Spiking Neural Networks: Concepts and Challenges. Front Neurosci 2022; 16:851774. [PMID: 35431782 PMCID: PMC9012531 DOI: 10.3389/fnins.2022.851774] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Accepted: 03/07/2022] [Indexed: 11/13/2022] Open
Abstract
Frequency-modulated continuous wave radar sensors play an essential role for assisted and autonomous driving as they are robust under all weather and light conditions. However, the rising number of transmitters and receivers for obtaining a higher angular resolution increases the cost for digital signal processing. One promising approach for energy-efficient signal processing is the usage of brain-inspired spiking neural networks (SNNs) implemented on neuromorphic hardware. In this article we perform a step-by-step analysis of automotive radar processing and argue how spiking neural networks could replace or complement the conventional processing. We provide SNN examples for two processing steps and evaluate their accuracy and computational efficiency. For radar target detection, an SNN with temporal coding is competitive to the conventional approach at a low compute overhead. Instead, our SNN for target classification achieves an accuracy close to a reference artificial neural network while requiring 200 times less operations. Finally, we discuss the specific requirements and challenges for SNN-based radar processing on neuromorphic hardware. This study proves the general applicability of SNNs for automotive radar processing and sustains the prospect of energy-efficient realizations in automated vehicles.
Collapse
Affiliation(s)
- Bernhard Vogginger
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
- *Correspondence: Bernhard Vogginger
| | - Felix Kreutz
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
- Infineon Technologies Dresden GmbH & Co., KG, Dresden, Germany
| | | | - Chen Liu
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
| | - Robin Dietrich
- Department of Informatics, Technical University of Munich, Munich, Germany
| | - Hector A. Gonzalez
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
| | - Daniel Scholz
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
- Infineon Technologies Dresden GmbH & Co., KG, Dresden, Germany
| | - Nico Reeb
- Department of Informatics, Technical University of Munich, Munich, Germany
| | - Daniel Auge
- Department of Informatics, Technical University of Munich, Munich, Germany
- Infineon Technologies AG, Munich, Germany
| | - Julian Hille
- Department of Informatics, Technical University of Munich, Munich, Germany
- Infineon Technologies AG, Munich, Germany
| | | | - Florian Mirus
- BMW Group, Research, New Technologies, Garching, Germany
| | | | - Alois Knoll
- Department of Informatics, Technical University of Munich, Munich, Germany
| | - Christian Mayr
- Chair of Highly-Parallel VLSI-Systems and Neuro-Microelectronics, Faculty of Electrical and Computer Engineering, Institute of Principles of Electrical and Electronic Engineering, Technische Universität Dresden, Dresden, Germany
- Centre for Tactile Internet (CeTI) With Human-In-The-Loop, Cluster of Excellence, Technische Universität Dresden, Dresden, Germany
| |
Collapse
|
22
|
Pehle C, Billaudelle S, Cramer B, Kaiser J, Schreiber K, Stradmann Y, Weis J, Leibfried A, Müller E, Schemmel J. The BrainScaleS-2 Accelerated Neuromorphic System With Hybrid Plasticity. Front Neurosci 2022; 16:795876. [PMID: 35281488 PMCID: PMC8907969 DOI: 10.3389/fnins.2022.795876] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2021] [Accepted: 01/27/2022] [Indexed: 12/30/2022] Open
Abstract
Since the beginning of information processing by electronic components, the nervous system has served as a metaphor for the organization of computational primitives. Brain-inspired computing today encompasses a class of approaches ranging from using novel nano-devices for computation to research into large-scale neuromorphic architectures, such as TrueNorth, SpiNNaker, BrainScaleS, Tianjic, and Loihi. While implementation details differ, spiking neural networks-sometimes referred to as the third generation of neural networks-are the common abstraction used to model computation with such systems. Here we describe the second generation of the BrainScaleS neuromorphic architecture, emphasizing applications enabled by this architecture. It combines a custom analog accelerator core supporting the accelerated physical emulation of bio-inspired spiking neural network primitives with a tightly coupled digital processor and a digital event-routing network.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | | | | | - Johannes Schemmel
- Electronic Visions, Kirchhoff-Institute for Physics, Heidelberg University, Heidelberg, Germany
| |
Collapse
|
23
|
Abstract
Neuromorphic systems aim to accomplish efficient computation in electronics by mirroring neurobiological principles. Taking advantage of neuromorphic technologies requires effective learning algorithms capable of instantiating high-performing neural networks, while also dealing with inevitable manufacturing variations of individual components, such as memristors or analog neurons. We present a learning framework resulting in bioinspired spiking neural networks with high performance, low inference latency, and sparse spike-coding schemes, which also self-corrects for device mismatch. We validate our approach on the BrainScaleS-2 analog spiking neuromorphic system, demonstrating state-of-the-art accuracy, low latency, and energy efficiency. Our work sketches a path for building powerful neuromorphic processors that take advantage of emerging analog technologies. To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum, but communicate with spikes, binary events in time. Analog neuromorphic hardware uses the same principles to emulate spiking neural networks with exceptional energy efficiency. However, instantiating high-performing spiking networks on such hardware remains a significant challenge due to device mismatch and the lack of efficient training algorithms. Surrogate gradient learning has emerged as a promising training strategy for spiking networks, but its applicability for analog neuromorphic systems has not been demonstrated. Here, we demonstrate surrogate gradient learning on the BrainScaleS-2 analog neuromorphic system using an in-the-loop approach. We show that learning self-corrects for device mismatch, resulting in competitive spiking network performance on both vision and speech benchmarks. Our networks display sparse spiking activity with, on average, less than one spike per hidden neuron and input, perform inference at rates of up to 85,000 frames per second, and consume less than 200 mW. In summary, our work sets several benchmarks for low-energy spiking network processing on analog neuromorphic hardware and paves the way for future on-chip learning algorithms.
Collapse
|
24
|
|