1
|
Wu K, E S, Yang N, Zhang A, Yan X, Mu C, Song Y. A novel approach to enhancing biomedical signal recognition via hybrid high-order information bottleneck driven spiking neural networks. Neural Netw 2025; 183:106976. [PMID: 39644595 DOI: 10.1016/j.neunet.2024.106976] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Revised: 10/16/2024] [Accepted: 11/06/2024] [Indexed: 12/09/2024]
Abstract
Biomedical signals, encapsulating vital physiological information, are pivotal in elucidating human traits and conditions, serving as a cornerstone for advancing human-machine interfaces. Nonetheless, the fidelity of biomedical signal interpretation is frequently compromised by pervasive noise sources such as skin, motion, and equipment interference, posing formidable challenges to precision recognition tasks. Concurrently, the burgeoning adoption of intelligent wearable devices illuminates a societal shift towards enhancing life and work through technological integration. This surge in popularity underscores the imperative for efficient, noise-resilient biomedical signal recognition methodologies, a quest that is both challenging and profoundly impactful. This study proposes a novel approach to enhancing biomedical signal recognition. The proposed approach employs a hierarchical information bottleneck mechanism within SNNs, quantifying the mutual information in different orders based on the depth of information flow in the network. Subsequently, these mutual information, together with the network's output and category labels, are restructured based on information theory principles to form the loss function used for training. A series of theoretical analyses and substantial experimental results have shown that this method can effectively compress noise in the data, and on the premise of low computational cost, it can also significantly outperform its vanilla counterpart in terms of classification performance.
Collapse
Affiliation(s)
- Kunlun Wu
- School of Artificial Intelligence, Anhui University, Hefei, 237090, China
| | - Shunzhuo E
- Suzhou High School of Jiangsu Province, Suzhou, 215011, China
| | - Ning Yang
- State Key Laboratory of Analog and Mixed-Signal VLSI, Institute of Microelectronics, University of Macau, Taipa, 999078, Macau
| | - Anguo Zhang
- School of Artificial Intelligence, Anhui University, Hefei, 237090, China.
| | - Xiaorong Yan
- Department of Neurosurgery, First Affiliated Hospital of Fujian Medical University, Fuzhou, 350506, China.
| | - Chaoxu Mu
- School of Artificial Intelligence, Anhui University, Hefei, 237090, China; School of Electrical and Information Engineering, Tianjin University, Tianjin, 300072, China
| | - Yongduan Song
- School of Artificial Intelligence, Anhui University, Hefei, 237090, China; Chongqing Key Laboratory of Autonomous Systems, Institute of Artificial Intelligence, School of Automation, Chongqing University, Chongqing, 400044, China
| |
Collapse
|
2
|
Zhang T, Wang Q, Xu B. Self-Lateral Propagation Elevates Synaptic Modifications in Spiking Neural Networks for the Efficient Spatial and Temporal Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:15359-15371. [PMID: 37389999 DOI: 10.1109/tnnls.2023.3286458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/02/2023]
Abstract
The brain's mystery for efficient and intelligent computation hides in the neuronal encoding, functional circuits, and plasticity principles in natural neural networks. However, many plasticity principles have not been fully incorporated into artificial or spiking neural networks (SNNs). Here, we report that incorporating a novel feature of synaptic plasticity found in natural networks, whereby synaptic modifications self-propagate to nearby synapses, named self-lateral propagation (SLP), could further improve the accuracy of SNNs in three benchmark spatial and temporal classification tasks. The SLP contains lateral pre ( SLP pre ) and lateral post ( SLP post ) synaptic propagation, describing the spread of synaptic modifications among output synapses made by axon collaterals or among converging synapses on the postsynaptic neuron, respectively. The SLP is biologically plausible and can lead to a coordinated synaptic modification within layers that endow higher efficiency without losing much accuracy. Furthermore, the experimental results showed the impressive role of SLP in sharpening the normal distribution of synaptic weights and broadening the more uniform distribution of misclassified samples, which are both considered essential for understanding the learning convergence and network generalization of neural networks.
Collapse
|
3
|
Chen T, She C, Wang L, Duan S. Memristive leaky integrate-and-fire neuron and learnable straight-through estimator in spiking neural networks. Cogn Neurodyn 2024; 18:3075-3091. [PMID: 39555273 PMCID: PMC11564454 DOI: 10.1007/s11571-024-10133-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Revised: 04/28/2024] [Accepted: 05/15/2024] [Indexed: 11/19/2024] Open
Abstract
Compared to artificial neural networks (ANNs), spiking neural networks (SNNs) present a more biologically plausible model of neural system dynamics. They rely on sparse binary spikes to communicate information and operate in an asynchronous, event-driven manner. Despite the high heterogeneity of the neural system at the neuronal level, most current SNNs employ the widely used leaky integrate-and-fire (LIF) neuron model, which assumes uniform membrane-related parameters throughout the entire network. This approach hampers the expressiveness of spiking neurons and restricts the diversity of neural dynamics. In this paper, we propose replacing the resistor in the LIF model with a discrete memristor to obtain the heterogeneous memristive LIF (MLIF) model. The memristance of the discrete memristor is determined by the voltage and flux at its terminals, leading to dynamic changes in the membrane time parameter of the MLIF model. SNNs composed of MLIF neurons can not only learn synaptic weights but also adaptively change membrane time parameters according to the membrane potential of the neuron, enhancing the learning ability and expression of SNNs. Furthermore, since the proper threshold of spiking neurons can improve the information capacity of SNNs, a learnable straight-through estimator (LSTE) is proposed. The LSTE, based on the straight-through estimator (STE) surrogate function, features a learnable threshold that facilitates the backward propagation of gradients through neurons firing spikes. Extensive experiments on several popular static and neuromorphic benchmark datasets demonstrate the effectiveness of the proposed MLIF and LSTE, especially on the DVS-CIFAR10 dataset, where we achieved the top-1 accuracy of 84.40 % .
Collapse
Affiliation(s)
- Tao Chen
- College of Artificial Intelligence, Southwest University, Chongqing, 400715 China
| | - Chunyan She
- College of Artificial Intelligence, Southwest University, Chongqing, 400715 China
| | - Lidan Wang
- College of Artificial Intelligence, Southwest University, Chongqing, 400715 China
- National and Local Joint Engineering Research Center of Intelligent Transmission and Control Technology, Chongqing, 400715 China
- Chongqing Key Laboratory of Brain-inspired Computing and Intelligent Chips, Chongqing, 400715 China
| | - Shukai Duan
- College of Artificial Intelligence, Southwest University, Chongqing, 400715 China
- National and Local Joint Engineering Research Center of Intelligent Transmission and Control Technology, Chongqing, 400715 China
- Chongqing Key Laboratory of Brain-inspired Computing and Intelligent Chips, Chongqing, 400715 China
| |
Collapse
|
4
|
Goupy G, Tirilly P, Bilasco IM. Paired competing neurons improving STDP supervised local learning in Spiking Neural Networks. Front Neurosci 2024; 18:1401690. [PMID: 39119458 PMCID: PMC11307446 DOI: 10.3389/fnins.2024.1401690] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2024] [Accepted: 07/11/2024] [Indexed: 08/10/2024] Open
Abstract
Direct training of Spiking Neural Networks (SNNs) on neuromorphic hardware has the potential to significantly reduce the energy consumption of artificial neural network training. SNNs trained with Spike Timing-Dependent Plasticity (STDP) benefit from gradient-free and unsupervised local learning, which can be easily implemented on ultra-low-power neuromorphic hardware. However, classification tasks cannot be performed solely with unsupervised STDP. In this paper, we propose Stabilized Supervised STDP (S2-STDP), a supervised STDP learning rule to train the classification layer of an SNN equipped with unsupervised STDP for feature extraction. S2-STDP integrates error-modulated weight updates that align neuron spikes with desired timestamps derived from the average firing time within the layer. Then, we introduce a training architecture called Paired Competing Neurons (PCN) to further enhance the learning capabilities of our classification layer trained with S2-STDP. PCN associates each class with paired neurons and encourages neuron specialization toward target or non-target samples through intra-class competition. We evaluate our methods on image recognition datasets, including MNIST, Fashion-MNIST, and CIFAR-10. Results show that our methods outperform state-of-the-art supervised STDP learning rules, for comparable architectures and numbers of neurons. Further analysis demonstrates that the use of PCN enhances the performance of S2-STDP, regardless of the hyperparameter set and without introducing any additional hyperparameters.
Collapse
|
5
|
Tenzin S, Rassau A, Chai D. Application of Event Cameras and Neuromorphic Computing to VSLAM: A Survey. Biomimetics (Basel) 2024; 9:444. [PMID: 39056885 PMCID: PMC11274992 DOI: 10.3390/biomimetics9070444] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 07/15/2024] [Accepted: 07/18/2024] [Indexed: 07/28/2024] Open
Abstract
Simultaneous Localization and Mapping (SLAM) is a crucial function for most autonomous systems, allowing them to both navigate through and create maps of unfamiliar surroundings. Traditional Visual SLAM, also commonly known as VSLAM, relies on frame-based cameras and structured processing pipelines, which face challenges in dynamic or low-light environments. However, recent advancements in event camera technology and neuromorphic processing offer promising opportunities to overcome these limitations. Event cameras inspired by biological vision systems capture the scenes asynchronously, consuming minimal power but with higher temporal resolution. Neuromorphic processors, which are designed to mimic the parallel processing capabilities of the human brain, offer efficient computation for real-time data processing of event-based data streams. This paper provides a comprehensive overview of recent research efforts in integrating event cameras and neuromorphic processors into VSLAM systems. It discusses the principles behind event cameras and neuromorphic processors, highlighting their advantages over traditional sensing and processing methods. Furthermore, an in-depth survey was conducted on state-of-the-art approaches in event-based SLAM, including feature extraction, motion estimation, and map reconstruction techniques. Additionally, the integration of event cameras with neuromorphic processors, focusing on their synergistic benefits in terms of energy efficiency, robustness, and real-time performance, was explored. The paper also discusses the challenges and open research questions in this emerging field, such as sensor calibration, data fusion, and algorithmic development. Finally, the potential applications and future directions for event-based SLAM systems are outlined, ranging from robotics and autonomous vehicles to augmented reality.
Collapse
Affiliation(s)
| | - Alexander Rassau
- School of Engineering, Edith Cowan University, Perth, WA 6027, Australia; (S.T.); (D.C.)
| | | |
Collapse
|
6
|
Wang Y, Wang Y, Zhang X, Du J, Zhang T, Xu B. Brain topology improved spiking neural network for efficient reinforcement learning of continuous control. Front Neurosci 2024; 18:1325062. [PMID: 38694900 PMCID: PMC11062182 DOI: 10.3389/fnins.2024.1325062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 03/27/2024] [Indexed: 05/04/2024] Open
Abstract
The brain topology highly reflects the complex cognitive functions of the biological brain after million-years of evolution. Learning from these biological topologies is a smarter and easier way to achieve brain-like intelligence with features of efficiency, robustness, and flexibility. Here we proposed a brain topology-improved spiking neural network (BT-SNN) for efficient reinforcement learning. First, hundreds of biological topologies are generated and selected as subsets of the Allen mouse brain topology with the help of the Tanimoto hierarchical clustering algorithm, which has been widely used in analyzing key features of the brain connectome. Second, a few biological constraints are used to filter out three key topology candidates, including but not limited to the proportion of node functions (e.g., sensation, memory, and motor types) and network sparsity. Third, the network topology is integrated with the hybrid numerical solver-improved leaky-integrated and fire neurons. Fourth, the algorithm is then tuned with an evolutionary algorithm named adaptive random search instead of backpropagation to guide synaptic modifications without affecting raw key features of the topology. Fifth, under the test of four animal-survival-like RL tasks (i.e., dynamic controlling in Mujoco), the BT-SNN can achieve higher scores than not only counterpart SNN using random topology but also some classical ANNs (i.e., long-short-term memory and multi-layer perception). This result indicates that the research effort of incorporating biological topology and evolutionary learning rules has much in store for the future.
Collapse
Affiliation(s)
- Yongjian Wang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Yansong Wang
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Future Technology, University of Chinese Academy of Sciences, Beijing, China
| | - Xinhe Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Jiulin Du
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Life Science and Technology, ShanghaiTech University, Shanghai, China
| | - Tielin Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
| | - Bo Xu
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- Institute of Neuroscience, State Key Laboratory of Neuroscience, Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- School of Future Technology, University of Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
7
|
Bacho F, Chu D. Low-variance Forward Gradients using Direct Feedback Alignment and momentum. Neural Netw 2024; 169:572-583. [PMID: 37956574 DOI: 10.1016/j.neunet.2023.10.051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2023] [Revised: 10/24/2023] [Accepted: 10/30/2023] [Indexed: 11/15/2023]
Abstract
Supervised learning in deep neural networks is commonly performed using error backpropagation. However, the sequential propagation of errors during the backward pass limits its scalability and applicability to low-powered neuromorphic hardware. Therefore, there is growing interest in finding local alternatives to backpropagation. Recently proposed methods based on forward-mode automatic differentiation suffer from high variance in large deep neural networks, which affects convergence. In this paper, we propose the Forward Direct Feedback Alignment algorithm that combines Activity-Perturbed Forward Gradients with Direct Feedback Alignment and momentum. We provide both theoretical proofs and empirical evidence that our proposed method achieves lower variance than forward gradient techniques. In this way, our approach enables faster convergence and better performance when compared to other local alternatives to backpropagation and opens a new perspective for the development of online learning algorithms compatible with neuromorphic systems.
Collapse
Affiliation(s)
- Florian Bacho
- CEMS, School of Computing, University of Kent, Canterbury, United Kingdom.
| | - Dominique Chu
- CEMS, School of Computing, University of Kent, Canterbury, United Kingdom.
| |
Collapse
|
8
|
Zhang T, Cheng X, Jia S, Li CT, Poo MM, Xu B. A brain-inspired algorithm that mitigates catastrophic forgetting of artificial and spiking neural networks with low computational cost. SCIENCE ADVANCES 2023; 9:eadi2947. [PMID: 37624895 PMCID: PMC10456855 DOI: 10.1126/sciadv.adi2947] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/16/2023] [Accepted: 07/27/2023] [Indexed: 08/27/2023]
Abstract
Neuromodulators in the brain act globally at many forms of synaptic plasticity, represented as metaplasticity, which is rarely considered by existing spiking (SNNs) and nonspiking artificial neural networks (ANNs). Here, we report an efficient brain-inspired computing algorithm for SNNs and ANNs, referred to here as neuromodulation-assisted credit assignment (NACA), which uses expectation signals to induce defined levels of neuromodulators to selective synapses, whereby the long-term synaptic potentiation and depression are modified in a nonlinear manner depending on the neuromodulator level. The NACA algorithm achieved high recognition accuracy with substantially reduced computational cost in learning spatial and temporal classification tasks. Notably, NACA was also verified as efficient for learning five different class continuous learning tasks with varying degrees of complexity, exhibiting a markedly mitigated catastrophic forgetting at low computational cost. Mapping synaptic weight changes showed that these benefits could be explained by the sparse and targeted synaptic modifications attributed to expectation-based global neuromodulation.
Collapse
Affiliation(s)
- Tielin Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
- Shanghai Center for Brain Science and Brain-inspired Technology, Lingang Laboratory, Shanghai 200031, China
| | - Xiang Cheng
- Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Shuncheng Jia
- Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Chengyu T Li
- Shanghai Center for Brain Science and Brain-inspired Technology, Lingang Laboratory, Shanghai 200031, China
- Center for Excellence in Brain Science and Intelligence Technology, Institute of Neuroscience, Chinese Academy of Sciences, Shanghai 200031, China
| | - Mu-ming Poo
- Shanghai Center for Brain Science and Brain-inspired Technology, Lingang Laboratory, Shanghai 200031, China
- Center for Excellence in Brain Science and Intelligence Technology, Institute of Neuroscience, Chinese Academy of Sciences, Shanghai 200031, China
| | - Bo Xu
- Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| |
Collapse
|
9
|
Yuan Y, Zhu Y, Wang J, Li R, Xu X, Fang T, Huo H, Wan L, Li Q, Liu N, Yang S. Incorporating structural plasticity into self-organization recurrent networks for sequence learning. Front Neurosci 2023; 17:1224752. [PMID: 37592946 PMCID: PMC10427342 DOI: 10.3389/fnins.2023.1224752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 07/13/2023] [Indexed: 08/19/2023] Open
Abstract
Introduction Spiking neural networks (SNNs), inspired by biological neural networks, have received a surge of interest due to its temporal encoding. Biological neural networks are driven by multiple plasticities, including spike timing-dependent plasticity (STDP), structural plasticity, and homeostatic plasticity, making network connection patterns and weights to change continuously during the lifecycle. However, it is unclear how these plasticities interact to shape neural networks and affect neural signal processing. Method Here, we propose a reward-modulated self-organization recurrent network with structural plasticity (RSRN-SP) to investigate this issue. Specifically, RSRN-SP uses spikes to encode information, and incorporate multiple plasticities including reward-modulated spike timing-dependent plasticity (R-STDP), homeostatic plasticity, and structural plasticity. On the one hand, combined with homeostatic plasticity, R-STDP is presented to guide the updating of synaptic weights. On the other hand, structural plasticity is utilized to simulate the growth and pruning of synaptic connections. Results and discussion Extensive experiments for sequential learning tasks are conducted to demonstrate the representational ability of the RSRN-SP, including counting task, motion prediction, and motion generation. Furthermore, the simulations also indicate that the characteristics arose from the RSRN-SP are consistent with biological observations.
Collapse
Affiliation(s)
- Ye Yuan
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Yongtong Zhu
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Jiaqi Wang
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Ruoshi Li
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Xin Xu
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Tao Fang
- Automation of Department, Shanghai Jiao Tong University, Shanghai, China
| | - Hong Huo
- Automation of Department, Shanghai Jiao Tong University, Shanghai, China
| | - Lihong Wan
- Origin Dynamics Intelligent Robot Co., Ltd., Zhengzhou, China
| | - Qingdu Li
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Na Liu
- School of Health Science and Engineering, Institute of Machine Intelligence, University of Shanghai for Science and Technology, Shanghai, China
| | - Shiyan Yang
- Eco-Environmental Protection Institution, Shanghai Academy of Agricultural Sciences, Shanghai, China
| |
Collapse
|
10
|
Oh S, An J, Cho S, Yoon R, Min KS. Memristor Crossbar Circuits Implementing Equilibrium Propagation for On-Device Learning. MICROMACHINES 2023; 14:1367. [PMID: 37512678 PMCID: PMC10384638 DOI: 10.3390/mi14071367] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Revised: 05/22/2023] [Accepted: 07/01/2023] [Indexed: 07/30/2023]
Abstract
Equilibrium propagation (EP) has been proposed recently as a new neural network training algorithm based on a local learning concept, where only local information is used to calculate the weight update of the neural network. Despite the advantages of local learning, numerical iteration for solving the EP dynamic equations makes the EP algorithm less practical for realizing edge intelligence hardware. Some analog circuits have been suggested to solve the EP dynamic equations physically, not numerically, using the original EP algorithm. However, there are still a few problems in terms of circuit implementation: for example, the need for storing the free-phase solution and the lack of essential peripheral circuits for calculating and updating synaptic weights. Therefore, in this paper, a new analog circuit technique is proposed to realize the EP algorithm in practical and implementable hardware. This work has two major contributions in achieving this objective. First, the free-phase and nudge-phase solutions are calculated by the proposed analog circuits simultaneously, not at different times. With this process, analog voltage memories or digital memories with converting circuits between digital and analog domains for storing the free-phase solution temporarily can be eliminated in the proposed EP circuit. Second, a simple EP learning rule relying on a fixed amount of conductance change per programming pulse is newly proposed and implemented in peripheral circuits. The modified EP learning rule can make the weight update circuit practical and implementable without requiring the use of a complicated program verification scheme. The proposed memristor conductance update circuit is simulated and verified for training synaptic weights on memristor crossbars. The simulation results showed that the proposed EP circuit could be used for realizing on-device learning in edge intelligence hardware.
Collapse
Affiliation(s)
- Seokjin Oh
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Jiyong An
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Seungmyeong Cho
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Rina Yoon
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| | - Kyeong-Sik Min
- School of Electrical Engineering, Kookmin University, Seoul 02707, Republic of Korea
| |
Collapse
|
11
|
Jia S, Zhang T, Zuo R, Xu B. Explaining cocktail party effect and McGurk effect with a spiking neural network improved by Motif-topology. Front Neurosci 2023; 17:1132269. [PMID: 37021133 PMCID: PMC10067589 DOI: 10.3389/fnins.2023.1132269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Accepted: 03/03/2023] [Indexed: 04/07/2023] Open
Abstract
Network architectures and learning principles have been critical in developing complex cognitive capabilities in artificial neural networks (ANNs). Spiking neural networks (SNNs) are a subset of ANNs that incorporate additional biological features such as dynamic spiking neurons, biologically specified architectures, and efficient and useful paradigms. Here we focus more on network architectures in SNNs, such as the meta operator called 3-node network motifs, which is borrowed from the biological network. We proposed a Motif-topology improved SNN (M-SNN), which is further verified efficient in explaining key cognitive phenomenon such as the cocktail party effect (a typical noise-robust speech-recognition task) and McGurk effect (a typical multi-sensory integration task). For M-SNN, the Motif topology is obtained by integrating the spatial and temporal motifs. These spatial and temporal motifs are first generated from the pre-training of spatial (e.g., MNIST) and temporal (e.g., TIDigits) datasets, respectively, and then applied to the previously introduced two cognitive effect tasks. The experimental results showed a lower computational cost and higher accuracy and a better explanation of some key phenomena of these two effects, such as new concept generation and anti-background noise. This mesoscale network motifs topology has much room for the future.
Collapse
Affiliation(s)
- Shuncheng Jia
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Tielin Zhang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- *Correspondence: Tielin Zhang
| | - Ruichen Zuo
- School of Information and Electronics, Beijing Institute of Technology, Beijing, China
| | - Bo Xu
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China
- Bo Xu
| |
Collapse
|