1
|
Liu J, Jiang C, Wei H, Wang Z, Sun L, Zhang S, Ni Y, Qu S, Yang L, Xu W. Vertically Integrated Monolithic Neuromorphic Nanowire Device for Physiological Information Processing. Nano Lett 2024; 24:4336-4345. [PMID: 38567915 DOI: 10.1021/acs.nanolett.3c04391] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/18/2024]
Abstract
This study demonstrates the conceptual design and fabrication of a vertically integrated monolithic (VIM) neuromorphic device. The device comprises an n-type SnO2 nanowire bottom channel connected by a shared gate to a p-type P3HT nanowire top channel. This architecture establishes two distinct neural pathways with different response behaviors. The device generates excitatory and inhibitory postsynaptic currents, mimicking the corelease mechanism of bilingual synapses. To enhance the signal processing efficiency, we employed a bipolar spike encoding strategy to convert fluctuating sensory signals to spike trains containing positive and negative pulses. Utilizing the neuromorphic platform for synaptic processing, physiological signals featuring bidirectional fluctuations, including electrocardiogram and breathing signals, can be classified with an accuracy of over 90%. The VIM device holds considerable promise as a solution for developing highly integrated neuromorphic hardware for healthcare and edge intelligence applications.
Collapse
Affiliation(s)
- Junchi Liu
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Chengpeng Jiang
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Huanhuan Wei
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Institutes of Physical Science and Information Technology, School of Materials Science and Engineering, Key Laboratory of Structure and Functional Regulation of Hybrid Materials, Anhui University, Ministry of Education, Hefei 230601, China
| | - Zixian Wang
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Lin Sun
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Song Zhang
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Yao Ni
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Shangda Qu
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Lu Yang
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| | - Wentao Xu
- Institute of Photoelectronic Thin Film Devices and Technology, Key Laboratory of Photoelectronic Thin Film Devices and Technology of Tianjin, College of Electronic Information and Optical Engineering, Engineering Research Center of Thin Film Photoelectronic Technology of Ministry of Education, Smart Sensing Interdisciplinary Science Center, Nankai University, Tianjin 300350, China
- Shenzhen Research Institute of Nankai University, Shenzhen 518000, China
| |
Collapse
|
2
|
Shim SK, Jang YH, Han J, Jeon JW, Shin DH, Kim YR, Han JK, Woo KS, Lee SH, Cheong S, Kim J, Seo H, Shin J, Hwang CS. 2Memristor-1Capacitor Integrated Temporal Kernel for High-Dimensional Data Mapping. Small 2024:e2306585. [PMID: 38212281 DOI: 10.1002/smll.202306585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/02/2023] [Revised: 12/01/2023] [Indexed: 01/13/2024]
Abstract
Compact but precise feature-extracting ability is core to processing complex computational tasks in neuromorphic hardware. Physical reservoir computing (RC) offers a robust framework to map temporal data into a high-dimensional space using the time dynamics of a material system, such as a volatile memristor. However, conventional physical RC systems have limited dynamics for the given material properties, restricting the methods to increase their dimensionality. This study proposes an integrated temporal kernel composed of a 2-memristor and 1-capacitor (2M1C) using a W/HfO2 /TiN memristor and TiN/ZrO2 /Al2 O3 /ZrO2 /TiN capacitor to achieve higher dimensionality and tunable dynamics. The kernel elements are carefully designed and fabricated into an integrated array, of which performances are evaluated under diverse conditions. By optimizing the time dynamics of the 2M1C kernel, each memristor simultaneously extracts complementary information from input signals. The MNIST benchmark digit classification task achieves a high accuracy of 94.3% with a (196×10) single-layer network. Analog input mapping ability is tested with a Mackey-Glass time series prediction, and the system records a normalized root mean square error of 0.04 with a 20×1 readout network, the smallest readout network ever used for Mackey-Glass prediction in RC. These performances demonstrate its high potential for efficient temporal data analysis.
Collapse
Affiliation(s)
- Sung Keun Shim
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Yoon Ho Jang
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Janguk Han
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Jeong Woo Jeon
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Dong Hoon Shin
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Yeong Rok Kim
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Joon-Kyu Han
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Kyung Seok Woo
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Soo Hyung Lee
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Sunwoo Cheong
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Jaehyun Kim
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Haengha Seo
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Jonghoon Shin
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| | - Cheol Seong Hwang
- Department of Materials Science and Engineering and Inter-university Semiconductor Research Center, College of Engineering, Seoul National University, Seoul, 08826, Republic of Korea
| |
Collapse
|
3
|
Han J, Lee J, Kim Y, Kim YB, Yun S, Lee S, Yu J, Lee KJ, Myung H, Choi Y. 3D Neuromorphic Hardware with Single Thin-Film Transistor Synapses Over Single Thin-Body Transistor Neurons by Monolithic Vertical Integration. Adv Sci (Weinh) 2023; 10:e2302380. [PMID: 37712147 PMCID: PMC10602577 DOI: 10.1002/advs.202302380] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Revised: 07/24/2023] [Indexed: 09/16/2023]
Abstract
Neuromorphic hardware with a spiking neural network (SNN) can significantly enhance the energy efficiency for artificial intelligence (AI) functions owing to its event-driven and spatiotemporally sparse operations. However, an artificial neuron and synapse based on complex complementary metal-oxide-semiconductor (CMOS) circuits limit the scalability and energy efficiency of neuromorphic hardware. In this work, a neuromorphic module is demonstrated composed of synapses over neurons realized by monolithic vertical integration. The synapse at top is a single thin-film transistor (1TFT-synapse) made of poly-crystalline silicon film and the neuron at bottom is another single transistor (1T-neuron) made of single-crystalline silicon. Excimer laser annealing (ELA) is applied to activate dopants for the 1TFT-synapse at the top and rapid thermal annealing (RTA) is applied to do so for the 1T-neuron at the bottom. Internal electro-thermal annealing (ETA) via the generation of Joule heat is also used to enhance the endurance of the 1TFT-synapse without transferring heat to the 1T-neuron at the bottom. As neuromorphic vision sensing, classification of American Sign Language (ASL) is conducted with the fabricated neuromorphic module. Its classification accuracy on ASL is ≈92.3% even after 204 800 update pulses.
Collapse
Affiliation(s)
- Joon‐Kyu Han
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Jung‐Woo Lee
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
- SK Hynix Inc.Icheon17336Republic of Korea
| | - Yeeun Kim
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Young Bin Kim
- Department of Materials Science and EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Seong‐Yun Yun
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Sang‐Won Lee
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Ji‐Man Yu
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Keon Jae Lee
- Department of Materials Science and EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Hyun Myung
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| | - Yang‐Kyu Choi
- School of Electrical EngineeringKorea Advanced Institute of Science and Technology (KAIST)291 Daehak‐ro, Yuseong‐guDaejeon34141Republic of Korea
| |
Collapse
|
4
|
Xue J, Xie L, Chen F, Wu L, Tian Q, Zhou Y, Ying R, Liu P. EdgeMap: An Optimized Mapping Toolchain for Spiking Neural Network in Edge Computing. Sensors (Basel) 2023; 23:6548. [PMID: 37514842 PMCID: PMC10383546 DOI: 10.3390/s23146548] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/28/2023] [Revised: 07/13/2023] [Accepted: 07/18/2023] [Indexed: 07/30/2023]
Abstract
Spiking neural networks (SNNs) have attracted considerable attention as third-generation artificial neural networks, known for their powerful, intelligent features and energy-efficiency advantages. These characteristics render them ideally suited for edge computing scenarios. Nevertheless, the current mapping schemes for deploying SNNs onto neuromorphic hardware face limitations such as extended execution times, low throughput, and insufficient consideration of energy consumption and connectivity, which undermine their suitability for edge computing applications. To address these challenges, we introduce EdgeMap, an optimized mapping toolchain specifically designed for deploying SNNs onto edge devices without compromising performance. EdgeMap consists of two main stages. The first stage involves partitioning the SNN graph into small neuron clusters based on the streaming graph partition algorithm, with the sizes of neuron clusters limited by the physical neuron cores. In the subsequent mapping stage, we adopt a multi-objective optimization algorithm specifically geared towards mitigating energy costs and communication costs for efficient deployment. EdgeMap-evaluated across four typical SNN applications-substantially outperforms other state-of-the-art mapping schemes. The performance improvements include a reduction in average latency by up to 19.8%, energy consumption by 57%, and communication cost by 58%. Moreover, EdgeMap exhibits an impressive enhancement in execution time by a factor of 1225.44×, alongside a throughput increase of up to 4.02×. These results highlight EdgeMap's efficiency and effectiveness, emphasizing its utility for deploying SNN applications in edge computing scenarios.
Collapse
Affiliation(s)
- Jianwei Xue
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Lisheng Xie
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Faquan Chen
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Liangshun Wu
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Qingyang Tian
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Yifan Zhou
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Rendong Ying
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Peilin Liu
- School of Electronic and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| |
Collapse
|
5
|
Yu M, Xiang T, P. S, Chu KTN, Amornpaisannon B, Tavva Y, Miriyala VPK, Carlson TE. A TTFS-based energy and utilization efficient neuromorphic CNN accelerator. Front Neurosci 2023; 17:1121592. [PMID: 37214405 PMCID: PMC10198466 DOI: 10.3389/fnins.2023.1121592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2022] [Accepted: 04/10/2023] [Indexed: 05/24/2023] Open
Abstract
Spiking neural networks (SNNs), which are a form of neuromorphic, brain-inspired AI, have the potential to be a power-efficient alternative to artificial neural networks (ANNs). Spikes that occur in SNN systems, also known as activations, tend to be extremely sparse, and low in number. This minimizes the number of data accesses typically needed for processing. In addition, SNN systems are typically designed to use addition operations which consume much less energy than the typical multiply and accumulate operations used in DNN systems. The vast majority of neuromorphic hardware designs support rate-based SNNs, where the information is encoded by spike rates. Generally, rate-based SNNs can be inefficient as a large number of spikes will be transmitted and processed during inference. One coding scheme that has the potential to improve efficiency is the time-to-first-spike (TTFS) coding, where the information isn't presented through the frequency of spikes, but instead through the relative spike arrival time. In TTFS-based SNNs, each neuron can only spike once during the entire inference process, and this results in high sparsity. The activation sparsity of TTFS-based SNNs is higher than rate-based SNNs, but TTFS-based SNNs have yet to achieve the same accuracy as rate-based SNNs. In this work, we propose two key improvements for TTFS-based SNN systems: (1) a novel optimization algorithm to improve the accuracy of TTFS-based SNNs and (2) a novel hardware accelerator for TTFS-based SNNs that uses a scalable and low-power design. Our work in TTFS coding and training improves the accuracy of TTFS-based SNNs to achieve state-of-the-art results on the MNIST and Fashion-MNIST datasets. Meanwhile, our work reduces the power consumption by at least 2.4×, 25.9×, and 38.4× over the state-of-the-art neuromorphic hardware on MNIST, Fashion-MNIST, and CIFAR10, respectively.
Collapse
Affiliation(s)
- Miao Yu
- School of Computing, Department of Computer Science, National University of Singapore, Singapore, Singapore
| | - Tingting Xiang
- School of Computing, Department of Computer Science, National University of Singapore, Singapore, Singapore
| | - Srivatsa P.
- School of Interactive Computing, Georgia Institute of Technology, Atlanta, GA, United States
| | - Kyle Timothy Ng Chu
- Centre for Quantum Technologies, National University of Singapore, Singapore, Singapore
| | - Burin Amornpaisannon
- School of Computing, Department of Computer Science, National University of Singapore, Singapore, Singapore
| | - Yaswanth Tavva
- School of Computing, Department of Computer Science, National University of Singapore, Singapore, Singapore
| | | | - Trevor E. Carlson
- School of Computing, Department of Computer Science, National University of Singapore, Singapore, Singapore
| |
Collapse
|
6
|
Schmid D, Oess T, Neumann H. Listen to the Brain-Auditory Sound Source Localization in Neuromorphic Computing Architectures. Sensors (Basel) 2023; 23:s23094451. [PMID: 37177655 PMCID: PMC10181665 DOI: 10.3390/s23094451] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2023] [Revised: 04/23/2023] [Accepted: 04/27/2023] [Indexed: 05/15/2023]
Abstract
Conventional processing of sensory input often relies on uniform sampling leading to redundant information and unnecessary resource consumption throughout the entire processing pipeline. Neuromorphic computing challenges these conventions by mimicking biology and employing distributed event-based hardware. Based on the task of lateral auditory sound source localization (SSL), we propose a generic approach to map biologically inspired neural networks to neuromorphic hardware. First, we model the neural mechanisms of SSL based on the interaural level difference (ILD). Afterward, we identify generic computational motifs within the model and transform them into spike-based components. A hardware-specific step then implements them on neuromorphic hardware. We exemplify our approach by mapping the neural SSL model onto two platforms, namely the IBM TrueNorth Neurosynaptic System and SpiNNaker. Both implementations have been tested on synthetic and real-world data in terms of neural tunings and readout characteristics. For synthetic stimuli, both implementations provide a perfect readout (100% accuracy). Preliminary real-world experiments yield accuracies of 78% (TrueNorth) and 13% (SpiNNaker), RMSEs of 41∘ and 39∘, and MAEs of 18∘ and 29∘, respectively. Overall, the proposed mapping approach allows for the successful implementation of the same SSL model on two different neuromorphic architectures paving the way toward more hardware-independent neural SSL.
Collapse
Affiliation(s)
- Daniel Schmid
- Institute of Neural Information Processing, Ulm University, James-Franck-Ring, 89081 Ulm, Germany
| | - Timo Oess
- Bernstein Center Freiburg, University of Freiburg, Hansastr. 9a, 79104 Freiburg im Breisgau, Germany
| | - Heiko Neumann
- Institute of Neural Information Processing, Ulm University, James-Franck-Ring, 89081 Ulm, Germany
| |
Collapse
|
7
|
Gerasimov JY, Tu D, Hitaishi V, Harikesh PC, Yang CY, Abrahamsson T, Rad M, Donahue MJ, Ejneby MS, Berggren M, Forchheimer R, Fabiano S. A Biologically Interfaced Evolvable Organic Pattern Classifier. Adv Sci (Weinh) 2023; 10:e2207023. [PMID: 36935358 DOI: 10.1002/advs.202207023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Revised: 02/16/2023] [Indexed: 05/18/2023]
Abstract
Future brain-computer interfaces will require local and highly individualized signal processing of fully integrated electronic circuits within the nervous system and other living tissue. New devices will need to be developed that can receive data from a sensor array, process these data into meaningful information, and translate that information into a format that can be interpreted by living systems. Here, the first example of interfacing a hardware-based pattern classifier with a biological nerve is reported. The classifier implements the Widrow-Hoff learning algorithm on an array of evolvable organic electrochemical transistors (EOECTs). The EOECTs' channel conductance is modulated in situ by electropolymerizing the semiconductor material within the channel, allowing for low voltage operation, high reproducibility, and an improvement in state retention by two orders of magnitude over state-of-the-art OECT devices. The organic classifier is interfaced with a biological nerve using an organic electrochemical spiking neuron to translate the classifier's output to a simulated action potential. The latter is then used to stimulate muscle contraction selectively based on the input pattern, thus paving the way for the development of adaptive neural interfaces for closed-loop therapeutic systems.
Collapse
Affiliation(s)
- Jennifer Y Gerasimov
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Deyu Tu
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Vivek Hitaishi
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Padinhare Cholakkal Harikesh
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Chi-Yuan Yang
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Tobias Abrahamsson
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Meysam Rad
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Mary J Donahue
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Malin Silverå Ejneby
- Department of Biomedical Engineering, Linköping University, Linköping, SE-581 83, Sweden
| | - Magnus Berggren
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| | - Robert Forchheimer
- Department of Electrical Engineering, Linköping University, Linköping, SE-581 83, Sweden
| | - Simone Fabiano
- Laboratory of Organic Electronics, Department of Science and Technology, Linköping University, Norrköping, SE-60174, Sweden
| |
Collapse
|
8
|
Skatchkovsky N, Jang H, Simeone O. Bayesian continual learning via spiking neural networks. Front Comput Neurosci 2022; 16:1037976. [PMID: 36465962 PMCID: PMC9708898 DOI: 10.3389/fncom.2022.1037976] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Accepted: 10/26/2022] [Indexed: 09/19/2023] Open
Abstract
Among the main features of biological intelligence are energy efficiency, capacity for continual adaptation, and risk management via uncertainty quantification. Neuromorphic engineering has been thus far mostly driven by the goal of implementing energy-efficient machines that take inspiration from the time-based computing paradigm of biological brains. In this paper, we take steps toward the design of neuromorphic systems that are capable of adaptation to changing learning tasks, while producing well-calibrated uncertainty quantification estimates. To this end, we derive online learning rules for spiking neural networks (SNNs) within a Bayesian continual learning framework. In it, each synaptic weight is represented by parameters that quantify the current epistemic uncertainty resulting from prior knowledge and observed data. The proposed online rules update the distribution parameters in a streaming fashion as data are observed. We instantiate the proposed approach for both real-valued and binary synaptic weights. Experimental results using Intel's Lava platform show the merits of Bayesian over frequentist learning in terms of capacity for adaptation and uncertainty quantification.
Collapse
Affiliation(s)
- Nicolas Skatchkovsky
- King's Communication, Learning and Information Processing (KCLIP) Lab, Department of Engineering, King's College London, London, United Kingdom
| | - Hyeryung Jang
- Department of Artificial Intelligence, Dongguk University, Seoul, South Korea
| | - Osvaldo Simeone
- King's Communication, Learning and Information Processing (KCLIP) Lab, Department of Engineering, King's College London, London, United Kingdom
| |
Collapse
|
9
|
Müller-Cleve SF, Fra V, Khacef L, Pequeño-Zurro A, Klepatsch D, Forno E, Ivanovich DG, Rastogi S, Urgese G, Zenke F, Bartolozzi C. Braille letter reading: A benchmark for spatio-temporal pattern recognition on neuromorphic hardware. Front Neurosci 2022; 16:951164. [PMID: 36440280 PMCID: PMC9695069 DOI: 10.3389/fnins.2022.951164] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2022] [Accepted: 10/19/2022] [Indexed: 03/25/2024] Open
Abstract
Spatio-temporal pattern recognition is a fundamental ability of the brain which is required for numerous real-world activities. Recent deep learning approaches have reached outstanding accuracies in such tasks, but their implementation on conventional embedded solutions is still very computationally and energy expensive. Tactile sensing in robotic applications is a representative example where real-time processing and energy efficiency are required. Following a brain-inspired computing approach, we propose a new benchmark for spatio-temporal tactile pattern recognition at the edge through Braille letter reading. We recorded a new Braille letters dataset based on the capacitive tactile sensors of the iCub robot's fingertip. We then investigated the importance of spatial and temporal information as well as the impact of event-based encoding on spike-based computation. Afterward, we trained and compared feedforward and recurrent Spiking Neural Networks (SNNs) offline using Backpropagation Through Time (BPTT) with surrogate gradients, then we deployed them on the Intel Loihi neuromorphic chip for fast and efficient inference. We compared our approach to standard classifiers, in particular to the Long Short-Term Memory (LSTM) deployed on the embedded NVIDIA Jetson GPU, in terms of classification accuracy, power, and energy consumption together with computational delay. Our results show that the LSTM reaches ~97% of accuracy, outperforming the recurrent SNN by ~17% when using continuous frame-based data instead of event-based inputs. However, the recurrent SNN on Loihi with event-based inputs is ~500 times more energy-efficient than the LSTM on Jetson, requiring a total power of only ~30 mW. This work proposes a new benchmark for tactile sensing and highlights the challenges and opportunities of event-based encoding, neuromorphic hardware, and spike-based computing for spatio-temporal pattern recognition at the edge.
Collapse
Affiliation(s)
| | - Vittorio Fra
- Politecnico di Torino, Electronic Design Automation (EDA) Group, Torino, Italy
| | - Lyes Khacef
- Bio-Inspired Circuits and Systems Lab, Zernike Institute for Advanced Materials, Groningen Cognitive Systems and Materials Center, University of Groningen, Groningen, Netherlands
| | | | - Daniel Klepatsch
- Silicon Austria Labs, Johannes Kepler Universität (JKU) Linz Institute of Technology (LIT) Silicon Austria Labs (SAL) embedded Signal Processing and Machine Learning (eSPML) Lab, Graz, Austria
- Johannes Kepler Universität (JKU) Linz Institute of Technology (LIT) Silicon Austria Labs (SAL) embedded Signal Processing and Machine Learning (eSPML) Lab, Johannes Kepler University Linz, Graz, Austria
| | - Evelina Forno
- Politecnico di Torino, Electronic Design Automation (EDA) Group, Torino, Italy
| | - Diego G. Ivanovich
- Silicon Austria Labs, Johannes Kepler Universität (JKU) Linz Institute of Technology (LIT) Silicon Austria Labs (SAL) embedded Signal Processing and Machine Learning (eSPML) Lab, Graz, Austria
- Johannes Kepler Universität (JKU) Linz Institute of Technology (LIT) Silicon Austria Labs (SAL) embedded Signal Processing and Machine Learning (eSPML) Lab, Johannes Kepler University Linz, Graz, Austria
| | - Shavika Rastogi
- International Centre for Neuromorphic Systems, Western Sydney University, Penrith, NSW, Australia
- Biocomputation Research Group, University of Hertfordshire, Hatfield, United Kingdom
| | - Gianvito Urgese
- Politecnico di Torino, Electronic Design Automation (EDA) Group, Torino, Italy
| | - Friedemann Zenke
- Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland
| | - Chiara Bartolozzi
- Istituto Italiano di Tecnologia, Event-Driven Perception in Robotics, Genoa, Italy
| |
Collapse
|
10
|
Kim M, Rehman MA, Lee D, Wang Y, Lim DH, Khan MF, Choi H, Shao QY, Suh J, Lee HS, Park HH. Filamentary and Interface-Type Memristors Based on Tantalum Oxide for Energy-Efficient Neuromorphic Hardware. ACS Appl Mater Interfaces 2022; 14:44561-44571. [PMID: 36164762 DOI: 10.1021/acsami.2c12296] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
To implement artificial neural networks (ANNs) based on memristor devices, it is essential to secure the linearity and symmetry in weight update characteristics of the memristor, and reliability in the cycle-to-cycle and device-to-device variations. This study experimentally demonstrated and compared the filamentary and interface-type resistive switching (RS) behaviors of tantalum oxide (Ta2O5 and TaO2)-based devices grown by atomic layer deposition (ALD) to propose a suitable RS type in terms of reliability and weight update characteristics. Although Ta2O5 is a strong candidate for memristor, the filament-type RS behavior of Ta2O5 does not fit well with ANNs demanding analog memory characteristics. Therefore, this study newly designed an interface-type TaO2 memristor and compared it to a filament type of Ta2O5 memristor to secure the weight update characteristics and reliability. The TaO2-based interface-type memristor exhibited gradual RS characteristics and area dependency in both high- and low-resistance states. In addition, compared to the filamentary memristor, the RS behaviors of the TaO2-based interface-type device exhibited higher suitability for the neuromorphic, symmetric, and linear long-term potentiation (LTP) and long-term depression (LTD). These findings suggest better types of memristors for implementing ionic memristor-based ANNs among the two types of RS mechanisms.
Collapse
Affiliation(s)
- Minjae Kim
- Department of Materials Science and Engineering, Yonsei University, Seoul 03722, South Korea
| | - Malik Abdul Rehman
- Department of Materials Science and Engineering, Yonsei University, Seoul 03722, South Korea
| | - Donghyun Lee
- Department of Materials Science and Engineering, Ulsan National Institute of Science and Technology, Ulsan 44919, Korea
| | - Yue Wang
- Department of Materials Science and Engineering, Yonsei University, Seoul 03722, South Korea
| | - Dong-Hyeok Lim
- Department of Materials Science and Engineering, Ulsan National Institute of Science and Technology, Ulsan 44919, Korea
| | - Muhammad Farooq Khan
- Department of Electrical Engineering, Sejong University, Seoul 05006, South Korea
| | - Haryeong Choi
- Department of Materials Science and Engineering, Yonsei University, Seoul 03722, South Korea
| | - Qing Yi Shao
- Provincial Key Laboratory of Nuclear Science, Institute of Quantum Matter, South China Normal University, Guangzhou 510006, China
| | - Joonki Suh
- Department of Materials Science and Engineering, Ulsan National Institute of Science and Technology, Ulsan 44919, Korea
| | - Hong-Sub Lee
- Department of Advanced Materials Engineering for Information and Electronics, Kyung Hee University, Yongin, Gyeonggi-do 17104, Korea
| | - Hyung-Ho Park
- Department of Materials Science and Engineering, Yonsei University, Seoul 03722, South Korea
| |
Collapse
|
11
|
Cramer B, Billaudelle S, Kanya S, Leibfried A, Grübl A, Karasenko V, Pehle C, Schreiber K, Stradmann Y, Weis J, Schemmel J, Zenke F. Surrogate gradients for analog neuromorphic computing. Proc Natl Acad Sci U S A 2022; 119:e2109194119. [PMID: 35042792 DOI: 10.1073/pnas.2109194119] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/25/2021] [Indexed: 11/18/2022] Open
Abstract
Neuromorphic systems aim to accomplish efficient computation in electronics by mirroring neurobiological principles. Taking advantage of neuromorphic technologies requires effective learning algorithms capable of instantiating high-performing neural networks, while also dealing with inevitable manufacturing variations of individual components, such as memristors or analog neurons. We present a learning framework resulting in bioinspired spiking neural networks with high performance, low inference latency, and sparse spike-coding schemes, which also self-corrects for device mismatch. We validate our approach on the BrainScaleS-2 analog spiking neuromorphic system, demonstrating state-of-the-art accuracy, low latency, and energy efficiency. Our work sketches a path for building powerful neuromorphic processors that take advantage of emerging analog technologies. To rapidly process temporal information at a low metabolic cost, biological neurons integrate inputs as an analog sum, but communicate with spikes, binary events in time. Analog neuromorphic hardware uses the same principles to emulate spiking neural networks with exceptional energy efficiency. However, instantiating high-performing spiking networks on such hardware remains a significant challenge due to device mismatch and the lack of efficient training algorithms. Surrogate gradient learning has emerged as a promising training strategy for spiking networks, but its applicability for analog neuromorphic systems has not been demonstrated. Here, we demonstrate surrogate gradient learning on the BrainScaleS-2 analog neuromorphic system using an in-the-loop approach. We show that learning self-corrects for device mismatch, resulting in competitive spiking network performance on both vision and speech benchmarks. Our networks display sparse spiking activity with, on average, less than one spike per hidden neuron and input, perform inference at rates of up to 85,000 frames per second, and consume less than 200 mW. In summary, our work sets several benchmarks for low-energy spiking network processing on analog neuromorphic hardware and paves the way for future on-chip learning algorithms.
Collapse
|
12
|
Zou C, Cui X, Kuang Y, Liu K, Wang Y, Wang X, Huang R. A Scatter-and-Gather Spiking Convolutional Neural Network on a Reconfigurable Neuromorphic Hardware. Front Neurosci 2021; 15:694170. [PMID: 34867142 PMCID: PMC8636746 DOI: 10.3389/fnins.2021.694170] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2021] [Accepted: 10/15/2021] [Indexed: 11/13/2022] Open
Abstract
Artificial neural networks (ANNs), like convolutional neural networks (CNNs), have achieved the state-of-the-art results for many machine learning tasks. However, inference with large-scale full-precision CNNs must cause substantial energy consumption and memory occupation, which seriously hinders their deployment on mobile and embedded systems. Highly inspired from biological brain, spiking neural networks (SNNs) are emerging as new solutions because of natural superiority in brain-like learning and great energy efficiency with event-driven communication and computation. Nevertheless, training a deep SNN remains a main challenge and there is usually a big accuracy gap between ANNs and SNNs. In this paper, we introduce a hardware-friendly conversion algorithm called "scatter-and-gather" to convert quantized ANNs to lossless SNNs, where neurons are connected with ternary {-1,0,1} synaptic weights. Each spiking neuron is stateless and more like original McCulloch and Pitts model, because it fires at most one spike and need be reset at each time step. Furthermore, we develop an incremental mapping framework to demonstrate efficient network deployments on a reconfigurable neuromorphic chip. Experimental results show our spiking LeNet on MNIST and VGG-Net on CIFAR-10 datasetobtain 99.37% and 91.91% classification accuracy, respectively. Besides, the presented mapping algorithm manages network deployment on our neuromorphic chip with maximum resource efficiency and excellent flexibility. Our four-spike LeNet and VGG-Net on chip can achieve respective real-time inference speed of 0.38 ms/image, 3.24 ms/image, and an average power consumption of 0.28 mJ/image and 2.3 mJ/image at 0.9 V, 252 MHz, which is nearly two orders of magnitude more efficient than traditional GPUs.
Collapse
Affiliation(s)
- Chenglong Zou
- Institute of Microelectronics, Peking University, Beijing, China.,School of ECE, Peking University Shenzhen Graduate School, Shenzhen, China
| | - Xiaoxin Cui
- Institute of Microelectronics, Peking University, Beijing, China
| | - Yisong Kuang
- Institute of Microelectronics, Peking University, Beijing, China
| | - Kefei Liu
- Institute of Microelectronics, Peking University, Beijing, China
| | - Yuan Wang
- Institute of Microelectronics, Peking University, Beijing, China
| | - Xinan Wang
- School of ECE, Peking University Shenzhen Graduate School, Shenzhen, China
| | - Ru Huang
- Institute of Microelectronics, Peking University, Beijing, China
| |
Collapse
|
13
|
Abstract
The promise of neuromorphic computing to develop ultra-low-power intelligent devices lies in its ability to localize information processing and memory storage in synaptic circuits much like the synapses in the brain. Spiking neural networks modeled using high-resolution synapses and armed with local unsupervised learning rules like spike time-dependent plasticity (STDP) have shown promising results in tasks such as pattern detection and image classification. However, designing and implementing a conventional, multibit STDP circuit becomes complex both in terms of the circuitry and the required silicon area. In this work, we introduce a modified and hardware-friendly STDP learning (named adaptive STDP) implemented using just 4-bit synapses. We demonstrate the capability of this learning rule in a pattern recognition task, in which a neuron learns to recognize a specific spike pattern embedded within noisy inhomogeneous Poisson spikes. Our results demonstrate that the performance of the proposed learning rule (94% using just 4-bit synapses) is similar to the conventional STDP learning (96% using 64-bit floating-point precision). The models used in this study are ideal ones for a CMOS neuromorphic circuit with analog soma and synapse circuits and mixed-signal learning circuits. The learning circuit stores the synaptic weight in a 4-bit digital memory that is updated asynchronously. In circuit simulation with Taiwan Semiconductor Manufacturing Company (TSMC) 250 nm CMOS process design kit (PDK), the static power consumption of a single synapse and the energy per spike (to generate a synaptic current of amplitude 15 pA and time constant 3 ms) are less than 2 pW and 200 fJ, respectively. The static power consumption of the learning circuit is less than 135 pW, and the energy to process a pair of pre- and postsynaptic spikes corresponding to a single learning step is less than 235 pJ. A single 4-bit synapse (capable of being configured as excitatory, inhibitory, or shunting inhibitory) along with its learning circuitry and digital memory occupies around 17,250 μm2 of silicon area.
Collapse
Affiliation(s)
- Ashish Gautam
- Department of Electrical Engineering and Information Systems, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| | - Takashi Kohno
- Institute of Industrial Science, The University of Tokyo, Tokyo, Japan
| |
Collapse
|
14
|
Chundi PK, Wang D, Kim SJ, Yang M, Cerqueira JP, Kang J, Jung S, Kim S, Seok M. Always-On Sub-Microwatt Spiking Neural Network Based on Spike-Driven Clock- and Power-Gating for an Ultra-Low-Power Intelligent Device. Front Neurosci 2021; 15:684113. [PMID: 34354559 PMCID: PMC8329666 DOI: 10.3389/fnins.2021.684113] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2021] [Accepted: 06/21/2021] [Indexed: 11/13/2022] Open
Abstract
This paper presents a novel spiking neural network (SNN) classifier architecture for enabling always-on artificial intelligent (AI) functions, such as keyword spotting (KWS) and visual wake-up, in ultra-low-power internet-of-things (IoT) devices. Such always-on hardware tends to dominate the power efficiency of an IoT device and therefore it is paramount to minimize its power dissipation. A key observation is that the input signal to always-on hardware is typically sparse in time. This is a great opportunity that a SNN classifier can leverage because the switching activity and the power consumption of SNN hardware can scale with spike rate. To leverage this scalability, the proposed SNN classifier architecture employs event-driven architecture, especially fine-grained clock generation and gating and fine-grained power gating, to obtain very low static power dissipation. The prototype is fabricated in 65 nm CMOS and occupies an area of 1.99 mm2. At 0.52 V supply voltage, it consumes 75 nW at no input activity and less than 300 nW at 100% input activity. It still maintains competitive inference accuracy for KWS and other always-on classification workloads. The prototype achieved a power consumption reduction of over three orders of magnitude compared to the state-of-the-art for SNN hardware and of about 2.3X compared to the state-of-the-art KWS hardware.
Collapse
Affiliation(s)
- Pavan Kumar Chundi
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| | - Dewei Wang
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| | - Sung Justin Kim
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| | - Minhao Yang
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| | - Joao Pedro Cerqueira
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| | | | | | | | - Mingoo Seok
- Department of Electrical Engineering, Columbia University, New York City, NY, United States
| |
Collapse
|
15
|
Hulea M, Uleru GI, Caruntu CF. Adaptive SNN for Anthropomorphic Finger Control. Sensors (Basel) 2021; 21:2730. [PMID: 33924453 DOI: 10.3390/s21082730] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/18/2021] [Revised: 04/09/2021] [Accepted: 04/10/2021] [Indexed: 11/16/2022]
Abstract
Anthropomorphic hands that mimic the smoothness of human hand motions should be controlled by artificial units of high biological plausibility. Adaptability is among the characteristics of such control units, which provides the anthropomorphic hand with the ability to learn motions. This paper presents a simple structure of an adaptive spiking neural network implemented in analogue hardware that can be trained using Hebbian learning mechanisms to rotate the metacarpophalangeal joint of a robotic finger towards targeted angle intervals. Being bioinspired, the spiking neural network drives actuators made of shape memory alloy and receives feedback from neuromorphic sensors that convert the joint rotation angle and compression force into the spiking frequency. The adaptive SNN activates independent neural paths that correspond to angle intervals and learns in which of these intervals the rotation the finger rotation is stopped by an external force. Learning occurs when angle-specific neural paths are stimulated concurrently with the supraliminar stimulus that activates all the neurons that inhibit the SNN output stopping the finger. The results showed that after learning, the finger stopped in the angle interval in which the angle-specific neural path was active, without the activation of the supraliminar stimulus. The proposed concept can be used to implement control units for anthropomorphic robots that are able to learn motions unsupervised, based on principles of high biological plausibility.
Collapse
|
16
|
Abstract
Despite the success of Deep Neural Networks-a type of Artificial Neural Network (ANN)-in problem domains such as image recognition and speech processing, the energy and processing demands during both training and deployment are growing at an unsustainable rate in the push for greater accuracy. There is a temptation to look for radical new approaches to these applications, and one such approach is the notion that replacing the abstract neuron used in most deep networks with a more biologically-plausible spiking neuron might lead to savings in both energy and resource cost. The most common spiking networks use rate-coded neurons for which a simple translation from a pre-trained ANN to an equivalent spike-based network (SNN) is readily achievable. But does the spike-based network offer an improvement of energy efficiency over the original deep network? In this work, we consider the digital implementations of the core steps in an ANN and the equivalent steps in a rate-coded spiking neural network. We establish a simple method of assessing the relative advantages of rate-based spike encoding over a conventional ANN model. Assuming identical underlying silicon technology we show that most rate-coded spiking network implementations will not be more energy or resource efficient than the original ANN, concluding that more imaginative uses of spikes are required to displace conventional ANNs as the dominant computing framework for neural computation.
Collapse
|
17
|
He W, Huang J, Wang T, Lin Y, He J, Zhou X, Li P, Wang Y, Wu N, Shi C. A High-Speed Low-Cost VLSI System Capable of On-Chip Online Learning for Dynamic Vision Sensor Data Classification. Sensors (Basel) 2020; 20:E4715. [PMID: 32825560 DOI: 10.3390/s20174715] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/04/2020] [Revised: 08/09/2020] [Accepted: 08/16/2020] [Indexed: 11/21/2022]
Abstract
This paper proposes a high-speed low-cost VLSI system capable of on-chip online learning for classifying address-event representation (AER) streams from dynamic vision sensor (DVS) retina chips. The proposed system executes a lightweight statistic algorithm based on simple binary features extracted from AER streams and a Random Ferns classifier to classify these features. The proposed system’s characteristics of multi-level pipelines and parallel processing circuits achieves a high throughput up to 1 spike event per clock cycle for AER data processing. Thanks to the nature of the lightweight algorithm, our hardware system is realized in a low-cost memory-centric paradigm. In addition, the system is capable of on-chip online learning to flexibly adapt to different in-situ application scenarios. The extra overheads for on-chip learning in terms of time and resource consumption are quite low, as the training procedure of the Random Ferns is quite simple, requiring few auxiliary learning circuits. An FPGA prototype of the proposed VLSI system was implemented with 9.5~96.7% memory consumption and <11% computational and logic resources on a Xilinx Zynq-7045 chip platform. It was running at a clock frequency of 100 MHz and achieved a peak processing throughput up to 100 Meps (Mega events per second), with an estimated power consumption of 690 mW leading to a high energy efficiency of 145 Meps/W or 145 event/μJ. We tested the prototype system on MNIST-DVS, Poker-DVS, and Posture-DVS datasets, and obtained classification accuracies of 77.9%, 99.4% and 99.3%, respectively. Compared to prior works, our VLSI system achieves higher processing speeds, higher computing efficiency, comparable accuracy, and lower resource costs.
Collapse
|
18
|
Abstract
A growing body of work underlines striking similarities between biological neural networks and recurrent, binary neural networks. A relatively smaller body of work, however, addresses the similarities between learning dynamics employed in deep artificial neural networks and synaptic plasticity in spiking neural networks. The challenge preventing this is largely caused by the discrepancy between the dynamical properties of synaptic plasticity and the requirements for gradient backpropagation. Learning algorithms that approximate gradient backpropagation using local error functions can overcome this challenge. Here, we introduce Deep Continuous Local Learning (DECOLLE), a spiking neural network equipped with local error functions for online learning with no memory overhead for computing gradients. DECOLLE is capable of learning deep spatio temporal representations from spikes relying solely on local information, making it compatible with neurobiology and neuromorphic hardware. Synaptic plasticity rules are derived systematically from user-defined cost functions and neural dynamics by leveraging existing autodifferentiation methods of machine learning frameworks. We benchmark our approach on the event-based neuromorphic dataset N-MNIST and DvsGesture, on which DECOLLE performs comparably to the state-of-the-art. DECOLLE networks provide continuously learning machines that are relevant to biology and supportive of event-based, low-power computer vision architectures matching the accuracies of conventional computers on tasks where temporal precision and speed are essential.
Collapse
Affiliation(s)
- Jacques Kaiser
- FZI Research Center for Information Technology, Karlsruhe, Germany
| | - Hesham Mostafa
- Department of Bioengineering, University of California, San Diego, La Jolla, CA, United States
| | - Emre Neftci
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA, United States
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| |
Collapse
|
19
|
Liu K, Cui X, Zhong Y, Kuang Y, Wang Y, Tang H, Huang R. A Hardware Implementation of SNN-Based Spatio-Temporal Memory Model. Front Neurosci 2019; 13:835. [PMID: 31447641 PMCID: PMC6697024 DOI: 10.3389/fnins.2019.00835] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2019] [Accepted: 07/26/2019] [Indexed: 11/13/2022] Open
Abstract
Simulating human brain with hardware has been an attractive project for many years, since memory is one of the fundamental functions of our brains. Several memory models have been proposed up to now in order to unveil how the memory is organized in the brain. In this paper, we adopt spatio-temporal memory (STM) model, in which both associative memory and episodic memory are analyzed and emulated, as the reference of our hardware network architecture. Furthermore, some reasonable adaptations are carried out for the hardware implementation. We finally implement this memory model on FPGA, and additional experiments are performed to fine tune the parameters of our network deployed on FPGA.
Collapse
Affiliation(s)
- Kefei Liu
- Institute of Microelectronics, Peking University, Beijing, China
| | - Xiaoxin Cui
- Institute of Microelectronics, Peking University, Beijing, China.,National Key Laboratory of Science and Technology on Micro/Nano Fabrication, Peking University, Beijing, China
| | - Yi Zhong
- Institute of Microelectronics, Peking University, Beijing, China
| | - Yisong Kuang
- Institute of Microelectronics, Peking University, Beijing, China
| | - Yuan Wang
- Institute of Microelectronics, Peking University, Beijing, China.,National Key Laboratory of Science and Technology on Micro/Nano Fabrication, Peking University, Beijing, China
| | - Huajin Tang
- College of Computer Science, Sichuan University, Chengdu, China
| | - Ru Huang
- Institute of Microelectronics, Peking University, Beijing, China.,National Key Laboratory of Science and Technology on Micro/Nano Fabrication, Peking University, Beijing, China
| |
Collapse
|
20
|
Nguyen TV, Pham KV, Min KS. Hybrid Circuit of Memristor and Complementary Metal-Oxide-Semiconductor for Defect-Tolerant Spatial Pooling with Boost-Factor Adjustment. Materials (Basel) 2019; 12:E2122. [PMID: 31266255 DOI: 10.3390/ma12132122] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/17/2019] [Revised: 06/27/2019] [Accepted: 06/29/2019] [Indexed: 11/16/2022]
Abstract
Hierarchical Temporal Memory (HTM) has been known as a software framework to model the brain's neocortical operation. However, mimicking the brain's neocortical operation by not software but hardware is more desirable, because the hardware can not only describe the neocortical operation, but can also employ the brain's architectural advantages. To develop a hybrid circuit of memristor and Complementary Metal-Oxide-Semiconductor (CMOS) for realizing HTM's spatial pooler (SP) by hardware, memristor defects such as stuck-at-faults and variations should be considered. For solving the defect problem, we first show that the boost-factor adjustment can make HTM's SP defect-tolerant, because the false activation of defective columns are suppressed. Second, we propose a memristor-CMOS hybrid circuit with the boost-factor adjustment to realize this defect-tolerant SP by hardware. The proposed circuit does not rely on the conventional defect-aware mapping scheme, which cannot avoid the false activation of defective columns. For the Modified subset of National Institute of Standards and Technology (MNIST) vectors, the boost-factor adjusted crossbar with defects = 10% shows a rate loss of only ~0.6%, compared to the ideal crossbar with defects = 0%. On the contrary, the defect-aware mapping without the boost-factor adjustment demonstrates a significant rate loss of ~21.0%. The energy overhead of the boost-factor adjustment is only ~0.05% of the programming energy of memristor synapse crossbar.
Collapse
|
21
|
Abstract
Hyperparameters and learning algorithms for neuromorphic hardware are usually chosen by hand to suit a particular task. In contrast, networks of neurons in the brain were optimized through extensive evolutionary and developmental processes to work well on a range of computing and learning tasks. Occasionally this process has been emulated through genetic algorithms, but these require themselves hand-design of their details and tend to provide a limited range of improvements. We employ instead other powerful gradient-free optimization tools, such as cross-entropy methods and evolutionary strategies, in order to port the function of biological optimization processes to neuromorphic hardware. As an example, we show these optimization algorithms enable neuromorphic agents to learn very efficiently from rewards. In particular, meta-plasticity, i.e., the optimization of the learning rule which they use, substantially enhances reward-based learning capability of the hardware. In addition, we demonstrate for the first time Learning-to-Learn benefits from such hardware, in particular, the capability to extract abstract knowledge from prior learning experiences that speeds up the learning of new but related tasks. Learning-to-Learn is especially suited for accelerated neuromorphic hardware, since it makes it feasible to carry out the required very large number of network computations.
Collapse
Affiliation(s)
- Thomas Bohnstingl
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| | - Franz Scherr
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| | - Christian Pehle
- Kirchhoff-Institute for Physics, Ruprecht-Karls-Universität Heidelberg, Heidelberg, Germany
| | - Karlheinz Meier
- Kirchhoff-Institute for Physics, Ruprecht-Karls-Universität Heidelberg, Heidelberg, Germany
| | - Wolfgang Maass
- Institute for Theoretical Computer Science, Graz University of Technology, Graz, Austria
| |
Collapse
|
22
|
Stöckel A, Jenzen C, Thies M, Rückert U. Binary Associative Memories as a Benchmark for Spiking Neuromorphic Hardware. Front Comput Neurosci 2017; 11:71. [PMID: 28878642 PMCID: PMC5572441 DOI: 10.3389/fncom.2017.00071] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2016] [Accepted: 07/20/2017] [Indexed: 11/14/2022] Open
Abstract
Large-scale neuromorphic hardware platforms, specialized computer systems for energy efficient simulation of spiking neural networks, are being developed around the world, for example as part of the European Human Brain Project (HBP). Due to conceptual differences, a universal performance analysis of these systems in terms of runtime, accuracy and energy efficiency is non-trivial, yet indispensable for further hard- and software development. In this paper we describe a scalable benchmark based on a spiking neural network implementation of the binary neural associative memory. We treat neuromorphic hardware and software simulators as black-boxes and execute exactly the same network description across all devices. Experiments on the HBP platforms under varying configurations of the associative memory show that the presented method allows to test the quality of the neuron model implementation, and to explain significant deviations from the expected reference output.
Collapse
Affiliation(s)
- Andreas Stöckel
- Cognitronics and Sensor Systems, Cluster of Excellence Cognitive Interaction Technology, Faculty of Technology, Bielefeld UniversityBielefeld, Germany
| | - Christoph Jenzen
- Cognitronics and Sensor Systems, Cluster of Excellence Cognitive Interaction Technology, Faculty of Technology, Bielefeld UniversityBielefeld, Germany
| | - Michael Thies
- Cognitronics and Sensor Systems, Cluster of Excellence Cognitive Interaction Technology, Faculty of Technology, Bielefeld UniversityBielefeld, Germany
| | - Ulrich Rückert
- Cognitronics and Sensor Systems, Cluster of Excellence Cognitive Interaction Technology, Faculty of Technology, Bielefeld UniversityBielefeld, Germany
| |
Collapse
|
23
|
Abstract
Connecting biologically inspired neural simulations to physical or simulated embodiments can be useful both in robotics, for the development of a new kind of bio-inspired controllers, and in neuroscience, to test detailed brain models in complete action-perception loops. The aim of this work is to develop a fully spike-based, biologically inspired mechanism for the translation of proprioceptive feedback. The translation is achieved by implementing a computational model of neural activity of type Ia and type II afferent fibers of muscle spindles, the primary source of proprioceptive information, which, in mammals is regulated through fusimotor activation and provides necessary adjustments during voluntary muscle contractions. As such, both static and dynamic γ-motoneurons activities are taken into account in the proposed model. Information from the actual proprioceptive sensors (i.e., motor encoders) is then used to simulate the spindle contraction and relaxation, and therefore drive the neural activity. To assess the feasibility of this approach, the model is implemented on the NEST spiking neural network simulator and on the SpiNNaker neuromorphic hardware platform and tested on simulated and physical robotic platforms. The results demonstrate that the model can be used in both simulated and real-time robotic applications to translate encoder values into a biologically plausible neural activity. Thus, this model provides a completely spike-based building block, suitable for neuromorphic platforms, that will enable the development of sensory-motor closed loops which could include neural simulations of areas of the central nervous system or of low-level reflexes.
Collapse
Affiliation(s)
- Lorenzo Vannucci
- The BioRobotics Institute, Scuola Superiore Sant'AnnaPontedera, Italy
| | - Egidio Falotico
- The BioRobotics Institute, Scuola Superiore Sant'AnnaPontedera, Italy
| | - Cecilia Laschi
- The BioRobotics Institute, Scuola Superiore Sant'AnnaPontedera, Italy
| |
Collapse
|
24
|
Abstract
Network of neurons in the brain apply-unlike processors in our current generation of computer hardware-an event-based processing strategy, where short pulses (spikes) are emitted sparsely by neurons to signal the occurrence of an event at a particular point in time. Such spike-based computations promise to be substantially more power-efficient than traditional clocked processing schemes. However, it turns out to be surprisingly difficult to design networks of spiking neurons that can solve difficult computational problems on the level of single spikes, rather than rates of spikes. We present here a new method for designing networks of spiking neurons via an energy function. Furthermore, we show how the energy function of a network of stochastically firing neurons can be shaped in a transparent manner by composing the networks of simple stereotypical network motifs. We show that this design approach enables networks of spiking neurons to produce approximate solutions to difficult (NP-hard) constraint satisfaction problems from the domains of planning/optimization and verification/logical inference. The resulting networks employ noise as a computational resource. Nevertheless, the timing of spikes plays an essential role in their computations. Furthermore, networks of spiking neurons carry out for the Traveling Salesman Problem a more efficient stochastic search for good solutions compared with stochastic artificial neural networks (Boltzmann machines) and Gibbs sampling.
Collapse
Affiliation(s)
- Zeno Jonke
- Faculty of Computer Science and Biomedical Engineering, Institute for Theoretical Computer Science, Graz University of Technology Graz, Austria
| | - Stefan Habenschuss
- Faculty of Computer Science and Biomedical Engineering, Institute for Theoretical Computer Science, Graz University of Technology Graz, Austria
| | - Wolfgang Maass
- Faculty of Computer Science and Biomedical Engineering, Institute for Theoretical Computer Science, Graz University of Technology Graz, Austria
| |
Collapse
|
25
|
Joucla S, Ambroise M, Levi T, Lafon T, Chauvet P, Saïghi S, Bornat Y, Lewis N, Renaud S, Yvert B. Generation of Locomotor-Like Activity in the Isolated Rat Spinal Cord Using Intraspinal Electrical Microstimulation Driven by a Digital Neuromorphic CPG. Front Neurosci 2016; 10:67. [PMID: 27013936 PMCID: PMC4779903 DOI: 10.3389/fnins.2016.00067] [Citation(s) in RCA: 28] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2015] [Accepted: 02/15/2016] [Indexed: 01/02/2023] Open
Abstract
Neural prostheses based on electrical microstimulation offer promising perspectives to restore functions following lesions of the central nervous system (CNS). They require the identification of appropriate stimulation sites and the coordination of their activation to achieve the restoration of functional activity. On the long term, a challenging perspective is to control microstimulation by artificial neural networks hybridized to the living tissue. Regarding the use of this strategy to restore locomotor activity in the spinal cord, to date, there has been no proof of principle of such hybrid approach driving intraspinal microstimulation (ISMS). Here, we address a first step toward this goal in the neonatal rat spinal cord isolated ex vivo, which can display locomotor-like activity while offering an easy access to intraspinal circuitry. Microelectrode arrays were inserted in the lumbar region to determine appropriate stimulation sites to elicit elementary bursting patterns on bilateral L2/L5 ventral roots. Two intraspinal sites were identified at L1 level, one on each side of the spinal cord laterally from the midline and approximately at a median position dorso-ventrally. An artificial CPG implemented on digital integrated circuit (FPGA) was built to generate alternating activity and was hybridized to the living spinal cord to drive electrical microstimulation on these two identified sites. Using this strategy, sustained left-right and flexor-extensor alternating activity on bilateral L2/L5 ventral roots could be generated in either whole or thoracically transected spinal cords. These results are a first step toward hybrid artificial/biological solutions based on electrical microstimulation for the restoration of lost function in the injured CNS.
Collapse
Affiliation(s)
- Sébastien Joucla
- Centre National de la Recherche Scientifique, Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287Talence, France; Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287, University of BordeauxTalence, France
| | - Matthieu Ambroise
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Centre National de la Recherche Scientifique, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Timothée Levi
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Centre National de la Recherche Scientifique, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Thierry Lafon
- Centre National de la Recherche Scientifique, Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287Talence, France; Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287, University of BordeauxTalence, France
| | - Philippe Chauvet
- Centre National de la Recherche Scientifique, Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287Talence, France; Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287, University of BordeauxTalence, France
| | - Sylvain Saïghi
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Centre National de la Recherche Scientifique, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Yannick Bornat
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Bordeaux INP, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Noëlle Lewis
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Centre National de la Recherche Scientifique, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Sylvie Renaud
- Laboratoire de l'Intégration du Matériau au Système, UMR 5218, University of BordeauxTalence, France; Bordeaux INP, Laboratoire de l'Intégration du Matériau au Système, UMR 5218Talence, France
| | - Blaise Yvert
- Centre National de la Recherche Scientifique, Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287Talence, France; Institute for Cognitive and Integrative Neuroscience (INCIA), UMR 5287, University of BordeauxTalence, France; Institut National de la Santé et de la Recherche Médicale, Clinatec-Lab, U1205Grenoble, France; Université Grenoble Alpes, Clinatec-Lab, U1205Grenoble, France
| |
Collapse
|
26
|
Kudithipudi D, Saleh Q, Merkel C, Thesing J, Wysocki B. Design and Analysis of a Neuromemristive Reservoir Computing Architecture for Biosignal Processing. Front Neurosci 2016; 9:502. [PMID: 26869876 PMCID: PMC4740959 DOI: 10.3389/fnins.2015.00502] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2015] [Accepted: 12/18/2015] [Indexed: 11/13/2022] Open
Abstract
Reservoir computing (RC) is gaining traction in several signal processing domains, owing to its non-linear stateful computation, spatiotemporal encoding, and reduced training complexity over recurrent neural networks (RNNs). Previous studies have shown the effectiveness of software-based RCs for a wide spectrum of applications. A parallel body of work indicates that realizing RNN architectures using custom integrated circuits and reconfigurable hardware platforms yields significant improvements in power and latency. In this research, we propose a neuromemristive RC architecture, with doubly twisted toroidal structure, that is validated for biosignal processing applications. We exploit the device mismatch to implement the random weight distributions within the reservoir and propose mixed-signal subthreshold circuits for energy efficiency. A comprehensive analysis is performed to compare the efficiency of the neuromemristive RC architecture in both digital(reconfigurable) and subthreshold mixed-signal realizations. Both Electroencephalogram (EEG) and Electromyogram (EMG) biosignal benchmarks are used for validating the RC designs. The proposed RC architecture demonstrated an accuracy of 90 and 84% for epileptic seizure detection and EMG prosthetic finger control, respectively.
Collapse
Affiliation(s)
- Dhireesha Kudithipudi
- NanoComputing Research Laboratory, Department of Computer Engineering, Rochester Institute of Technology Rochester, NY, USA
| | - Qutaiba Saleh
- NanoComputing Research Laboratory, Department of Computer Engineering, Rochester Institute of Technology Rochester, NY, USA
| | - Cory Merkel
- NanoComputing Research Laboratory, Department of Computer Engineering, Rochester Institute of Technology Rochester, NY, USA
| | - James Thesing
- NanoComputing Research Laboratory, Department of Computer Engineering, Rochester Institute of Technology Rochester, NY, USA
| | - Bryant Wysocki
- Information Directorate, Air Force Research Laboratory Rome, NY, USA
| |
Collapse
|
27
|
Diamond A, Nowotny T, Schmuker M. Comparing Neuromorphic Solutions in Action: Implementing a Bio-Inspired Solution to a Benchmark Classification Task on Three Parallel-Computing Platforms. Front Neurosci 2016; 9:491. [PMID: 26778950 PMCID: PMC4705229 DOI: 10.3389/fnins.2015.00491] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2015] [Accepted: 12/10/2015] [Indexed: 01/24/2023] Open
Abstract
Neuromorphic computing employs models of neuronal circuits to solve computing problems. Neuromorphic hardware systems are now becoming more widely available and "neuromorphic algorithms" are being developed. As they are maturing toward deployment in general research environments, it becomes important to assess and compare them in the context of the applications they are meant to solve. This should encompass not just task performance, but also ease of implementation, speed of processing, scalability, and power efficiency. Here, we report our practical experience of implementing a bio-inspired, spiking network for multivariate classification on three different platforms: the hybrid digital/analog Spikey system, the digital spike-based SpiNNaker system, and GeNN, a meta-compiler for parallel GPU hardware. We assess performance using a standard hand-written digit classification task. We found that whilst a different implementation approach was required for each platform, classification performances remained in line. This suggests that all three implementations were able to exercise the model's ability to solve the task rather than exposing inherent platform limits, although differences emerged when capacity was approached. With respect to execution speed and power consumption, we found that for each platform a large fraction of the computing time was spent outside of the neuromorphic device, on the host machine. Time was spent in a range of combinations of preparing the model, encoding suitable input spiking data, shifting data, and decoding spike-encoded results. This is also where a large proportion of the total power was consumed, most markedly for the SpiNNaker and Spikey systems. We conclude that the simulation efficiency advantage of the assessed specialized hardware systems is easily lost in excessive host-device communication, or non-neuronal parts of the computation. These results emphasize the need to optimize the host-device communication architecture for scalability, maximum throughput, and minimum latency. Moreover, our results indicate that special attention should be paid to minimize host-device communication when designing and implementing networks for efficient neuromorphic computing.
Collapse
Affiliation(s)
- Alan Diamond
- School of Engineering and Informatics, University of SussexBrighton, UK
| | | | | |
Collapse
|
28
|
Abstract
Evaluating the effectiveness and performance of neuromorphic hardware is difficult. It is even more difficult when the task of interest is a closed-loop task; that is, a task where the output from the neuromorphic hardware affects some environment, which then in turn affects the hardware's future input. However, closed-loop situations are one of the primary potential uses of neuromorphic hardware. To address this, we present a methodology for generating closed-loop benchmarks that makes use of a hybrid of real physical embodiment and a type of “minimal” simulation. Minimal simulation has been shown to lead to robust real-world performance, while still maintaining the practical advantages of simulation, such as making it easy for the same benchmark to be used by many researchers. This method is flexible enough to allow researchers to explicitly modify the benchmarks to identify specific task domains where particular hardware excels. To demonstrate the method, we present a set of novel benchmarks that focus on motor control for an arbitrary system with unknown external forces. Using these benchmarks, we show that an error-driven learning rule can consistently improve motor control performance across a randomly generated family of closed-loop simulations, even when there are up to 15 interacting joints to be controlled.
Collapse
Affiliation(s)
- Terrence C Stewart
- Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| | - Travis DeWolf
- Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| | - Ashley Kleinhans
- Mobile Intelligent Autonomous Systems group, Council for Scientific and Industrial Research Pretoria, South Africa
| | - Chris Eliasmith
- Centre for Theoretical Neuroscience, University of Waterloo Waterloo, ON, Canada
| |
Collapse
|
29
|
Probst D, Petrovici MA, Bytschok I, Bill J, Pecevski D, Schemmel J, Meier K. Probabilistic inference in discrete spaces can be implemented into networks of LIF neurons. Front Comput Neurosci 2015; 9:13. [PMID: 25729361 PMCID: PMC4325917 DOI: 10.3389/fncom.2015.00013] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2014] [Accepted: 01/27/2015] [Indexed: 12/18/2022] Open
Abstract
The means by which cortical neural networks are able to efficiently solve inference problems remains an open question in computational neuroscience. Recently, abstract models of Bayesian computation in neural circuits have been proposed, but they lack a mechanistic interpretation at the single-cell level. In this article, we describe a complete theoretical framework for building networks of leaky integrate-and-fire neurons that can sample from arbitrary probability distributions over binary random variables. We test our framework for a model inference task based on a psychophysical phenomenon (the Knill-Kersten optical illusion) and further assess its performance when applied to randomly generated distributions. As the local computations performed by the network strongly depend on the interaction between neurons, we compare several types of couplings mediated by either single synapses or interneuron chains. Due to its robustness to substrate imperfections such as parameter noise and background noise correlations, our model is particularly interesting for implementation on novel, neuro-inspired computing architectures, which can thereby serve as a fast, low-power substrate for solving real-world inference problems.
Collapse
Affiliation(s)
- Dimitri Probst
- Kirchhoff Institute for Physics, University of HeidelbergHeidelberg, Germany
| | - Mihai A. Petrovici
- Kirchhoff Institute for Physics, University of HeidelbergHeidelberg, Germany
| | - Ilja Bytschok
- Kirchhoff Institute for Physics, University of HeidelbergHeidelberg, Germany
| | - Johannes Bill
- Institute for Theoretical Computer Science, Graz University of TechnologyGraz, Austria
| | - Dejan Pecevski
- Institute for Theoretical Computer Science, Graz University of TechnologyGraz, Austria
| | - Johannes Schemmel
- Kirchhoff Institute for Physics, University of HeidelbergHeidelberg, Germany
| | - Karlheinz Meier
- Kirchhoff Institute for Physics, University of HeidelbergHeidelberg, Germany
| |
Collapse
|
30
|
Galluppi F, Lagorce X, Stromatias E, Pfeiffer M, Plana LA, Furber SB, Benosman RB. A framework for plasticity implementation on the SpiNNaker neural architecture. Front Neurosci 2015; 8:429. [PMID: 25653580 PMCID: PMC4299433 DOI: 10.3389/fnins.2014.00429] [Citation(s) in RCA: 36] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2014] [Accepted: 12/07/2014] [Indexed: 11/21/2022] Open
Abstract
Many of the precise biological mechanisms of synaptic plasticity remain elusive, but simulations of neural networks have greatly enhanced our understanding of how specific global functions arise from the massively parallel computation of neurons and local Hebbian or spike-timing dependent plasticity rules. For simulating large portions of neural tissue, this has created an increasingly strong need for large scale simulations of plastic neural networks on special purpose hardware platforms, because synaptic transmissions and updates are badly matched to computing style supported by current architectures. Because of the great diversity of biological plasticity phenomena and the corresponding diversity of models, there is a great need for testing various hypotheses about plasticity before committing to one hardware implementation. Here we present a novel framework for investigating different plasticity approaches on the SpiNNaker distributed digital neural simulation platform. The key innovation of the proposed architecture is to exploit the reconfigurability of the ARM processors inside SpiNNaker, dedicating a subset of them exclusively to process synaptic plasticity updates, while the rest perform the usual neural and synaptic simulations. We demonstrate the flexibility of the proposed approach by showing the implementation of a variety of spike- and rate-based learning rules, including standard Spike-Timing dependent plasticity (STDP), voltage-dependent STDP, and the rate-based BCM rule. We analyze their performance and validate them by running classical learning experiments in real time on a 4-chip SpiNNaker board. The result is an efficient, modular, flexible and scalable framework, which provides a valuable tool for the fast and easy exploration of learning models of very different kinds on the parallel and reconfigurable SpiNNaker system.
Collapse
Affiliation(s)
- Francesco Galluppi
- Equipe de Vision et Calcul Naturel, Vision Institute, Université Pierre et Marie Curie, Unité Mixte de Recherche S968 Inserm, l'Université Pierre et Marie Curie, Centre National de la Recherche Scientifique Unité Mixte de Recherche 7210, Centre Hospitalier National d'Ophtalmologie des quinze-vingtsParis, France
| | - Xavier Lagorce
- Equipe de Vision et Calcul Naturel, Vision Institute, Université Pierre et Marie Curie, Unité Mixte de Recherche S968 Inserm, l'Université Pierre et Marie Curie, Centre National de la Recherche Scientifique Unité Mixte de Recherche 7210, Centre Hospitalier National d'Ophtalmologie des quinze-vingtsParis, France
| | - Evangelos Stromatias
- Advanced Processors Technology Group, School of Computer Science, University of ManchesterManchester, UK
| | - Michael Pfeiffer
- Institute of Neuroinformatics, University of Zürich and ETH ZürichZürich, Switzerland
| | - Luis A. Plana
- Advanced Processors Technology Group, School of Computer Science, University of ManchesterManchester, UK
| | - Steve B. Furber
- Advanced Processors Technology Group, School of Computer Science, University of ManchesterManchester, UK
| | - Ryad B. Benosman
- Equipe de Vision et Calcul Naturel, Vision Institute, Université Pierre et Marie Curie, Unité Mixte de Recherche S968 Inserm, l'Université Pierre et Marie Curie, Centre National de la Recherche Scientifique Unité Mixte de Recherche 7210, Centre Hospitalier National d'Ophtalmologie des quinze-vingtsParis, France
| |
Collapse
|
31
|
Affiliation(s)
- Corey M Thibeault
- Center for Neural and Emergent Systems, Information and System Sciences Laboratory, HRL Laboratories LLC. Malibu, CA, USA
| |
Collapse
|
32
|
Friedmann S, Frémaux N, Schemmel J, Gerstner W, Meier K. Reward-based learning under hardware constraints-using a RISC processor embedded in a neuromorphic substrate. Front Neurosci 2013; 7:160. [PMID: 24065877 PMCID: PMC3778319 DOI: 10.3389/fnins.2013.00160] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2013] [Accepted: 08/19/2013] [Indexed: 11/16/2022] Open
Abstract
In this study, we propose and analyze in simulations a new, highly flexible method of implementing synaptic plasticity in a wafer-scale, accelerated neuromorphic hardware system. The study focuses on globally modulated STDP, as a special use-case of this method. Flexibility is achieved by embedding a general-purpose processor dedicated to plasticity into the wafer. To evaluate the suitability of the proposed system, we use a reward modulated STDP rule in a spike train learning task. A single layer of neurons is trained to fire at specific points in time with only the reward as feedback. This model is simulated to measure its performance, i.e., the increase in received reward after learning. Using this performance as baseline, we then simulate the model with various constraints imposed by the proposed implementation and compare the performance. The simulated constraints include discretized synaptic weights, a restricted interface between analog synapses and embedded processor, and mismatch of analog circuits. We find that probabilistic updates can increase the performance of low-resolution weights, a simple interface between analog synapses and processor is sufficient for learning, and performance is insensitive to mismatch. Further, we consider communication latency between wafer and the conventional control computer system that is simulating the environment. This latency increases the delay, with which the reward is sent to the embedded processor. Because of the time continuous operation of the analog synapses, delay can cause a deviation of the updates as compared to the not delayed situation. We find that for highly accelerated systems latency has to be kept to a minimum. This study demonstrates the suitability of the proposed implementation to emulate the selected reward modulated STDP learning rule. It is therefore an ideal candidate for implementation in an upgraded version of the wafer-scale system developed within the BrainScaleS project.
Collapse
Affiliation(s)
- Simon Friedmann
- Kirchhoff Institute for Physics, Ruprecht-Karls-University Heidelberg Heidelberg, Germany
| | | | | | | | | |
Collapse
|
33
|
Pfeil T, Potjans TC, Schrader S, Potjans W, Schemmel J, Diesmann M, Meier K. Is a 4-bit synaptic weight resolution enough? - constraints on enabling spike-timing dependent plasticity in neuromorphic hardware. Front Neurosci 2012; 6:90. [PMID: 22822388 PMCID: PMC3398398 DOI: 10.3389/fnins.2012.00090] [Citation(s) in RCA: 68] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2012] [Accepted: 06/04/2012] [Indexed: 11/13/2022] Open
Abstract
Large-scale neuromorphic hardware systems typically bear the trade-off between detail level and required chip resources. Especially when implementing spike-timing dependent plasticity, reduction in resources leads to limitations as compared to floating point precision. By design, a natural modification that saves resources would be reducing synaptic weight resolution. In this study, we give an estimate for the impact of synaptic weight discretization on different levels, ranging from random walks of individual weights to computer simulations of spiking neural networks. The FACETS wafer-scale hardware system offers a 4-bit resolution of synaptic weights, which is shown to be sufficient within the scope of our network benchmark. Our findings indicate that increasing the resolution may not even be useful in light of further restrictions of customized mixed-signal synapses. In addition, variations due to production imperfections are investigated and shown to be uncritical in the context of the presented study. Our results represent a general framework for setting up and configuring hardware-constrained synapses. We suggest how weight discretization could be considered for other backends dedicated to large-scale simulations. Thus, our proposition of a good hardware verification practice may rise synergy effects between hardware developers and neuroscientists.
Collapse
Affiliation(s)
- Thomas Pfeil
- Kirchhoff Institute for Physics, Ruprecht-Karls-University Heidelberg Heidelberg, Germany
| | | | | | | | | | | | | |
Collapse
|
34
|
Bill J, Schuch K, Brüderle D, Schemmel J, Maass W, Meier K. Compensating Inhomogeneities of Neuromorphic VLSI Devices Via Short-Term Synaptic Plasticity. Front Comput Neurosci 2010; 4:129. [PMID: 21031027 PMCID: PMC2965017 DOI: 10.3389/fncom.2010.00129] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [What about the content of this article? (0)] [Affiliation(s)] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2010] [Accepted: 08/11/2010] [Indexed: 11/17/2022] Open
Abstract
Recent developments in neuromorphic hardware engineering make mixed-signal VLSI neural network models promising candidates for neuroscientific research tools and massively parallel computing devices, especially for tasks which exhaust the computing power of software simulations. Still, like all analog hardware systems, neuromorphic models suffer from a constricted configurability and production-related fluctuations of device characteristics. Since also future systems, involving ever-smaller structures, will inevitably exhibit such inhomogeneities on the unit level, self-regulation properties become a crucial requirement for their successful operation. By applying a cortically inspired self-adjusting network architecture, we show that the activity of generic spiking neural networks emulated on a neuromorphic hardware system can be kept within a biologically realistic firing regime and gain a remarkable robustness against transistor-level variations. As a first approach of this kind in engineering practice, the short-term synaptic depression and facilitation mechanisms implemented within an analog VLSI model of I&F neurons are functionally utilized for the purpose of network level stabilization. We present experimental data acquired both from the hardware model and from comparative software simulations which prove the applicability of the employed paradigm to neuromorphic VLSI devices.
Collapse
Affiliation(s)
- Johannes Bill
- Kirchhoff Institute for Physics, University of Heidelberg Heidelberg, Germany
| | | | | | | | | | | |
Collapse
|