1
|
Jia H, Tang S, Guo W, Pan P, Qian Y, Hu D, Dai Y, Yang Y, Geng C, Lv H. Differential diagnosis of congenital ventricular septal defect and atrial septal defect in children using deep learning-based analysis of chest radiographs. BMC Pediatr 2024; 24:661. [PMID: 39407181 PMCID: PMC11476512 DOI: 10.1186/s12887-024-05141-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Accepted: 10/09/2024] [Indexed: 10/20/2024] Open
Abstract
BACKGROUND Children with atrial septal defect (ASD) and ventricular septal defect (VSD) are frequently examined for respiratory symptoms, even when the underlying disease is not found. Chest radiographs often serve as the primary imaging modality. It is crucial to differentiate between ASD and VSD due to their distinct treatment. PURPOSE To assess whether deep learning analysis of chest radiographs can more effectively differentiate between ASD and VSD in children. METHODS In this retrospective study, chest radiographs and corresponding radiology reports from 1,194 patients were analyzed. The cases were categorized into a training set and a validation set, comprising 480 cases of ASD and 480 cases of VSD, and a test set with 115 cases of ASD and 119 cases of VSD. Four deep learning network models-ResNet-CBAM, InceptionV3, EfficientNet, and ViT-were developed for training, and a fivefold cross-validation method was employed to optimize the models. Receiver operating characteristic (ROC) curve analyses were conducted to assess the performance of each model. The most effective algorithm was compared with the interpretations provided by two radiologists on 234 images from the test group. RESULTS The average accuracy, sensitivity, and specificity of the four deep learning models in the differential diagnosis of VSD and ASD were higher than 70%. The AUC values of ResNet-CBAM, IncepetionV3, EfficientNet, and ViT were 0.87, 0.91, 0.90, and 0.66, respectively. Statistical analysis showed that the differential diagnosis efficiency of InceptionV3 was the highest, reaching 87% classification accuracy. The accuracy of InceptionV3 in the differential diagnosis of VSD and ASD was higher than that of the radiologists. CONCLUSIONS Deep learning methods such as IncepetionV3 based on chest radiographs in the study showed good performance for differential diagnosis of congenital VSD and ASD, which may be able to assist radiologists in diagnosis, education, and training, and reduce missed diagnosis and misdiagnosis.
Collapse
Affiliation(s)
- Huihui Jia
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Songqiao Tang
- School of Electronic & Information Engineering, Suzhou University of Science and Technology, 215009, Suzhou, China
| | - Wanliang Guo
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Peng Pan
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Yufeng Qian
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Dongliang Hu
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Yakang Dai
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, 215163, Suzhou, China
| | - Yang Yang
- Department of Radiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China
| | - Chen Geng
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, 215163, Suzhou, China.
- Jinan Guoke Medical Technology Development Co., Ltd, 250102, Shandong, China.
| | - Haitao Lv
- Department of Pediatric Cardiology, Children ' s Hospital of Soochow University, 215025, Suzhou, P. R. China.
| |
Collapse
|
2
|
Sun H, Liu R, Cai W, Wang J, Wang Y, Tang H, Cui Y, Yao D, Guo D. Reliable object tracking by multimodal hybrid feature extraction and transformer-based fusion. Neural Netw 2024; 178:106493. [PMID: 38970946 DOI: 10.1016/j.neunet.2024.106493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2024] [Revised: 05/21/2024] [Accepted: 06/25/2024] [Indexed: 07/08/2024]
Abstract
Visual object tracking, which is primarily based on visible light image sequences, encounters numerous challenges in complicated scenarios, such as low light conditions, high dynamic ranges, and background clutter. To address these challenges, incorporating the advantages of multiple visual modalities is a promising solution for achieving reliable object tracking. However, the existing approaches usually integrate multimodal inputs through adaptive local feature interactions, which cannot leverage the full potential of visual cues, thus resulting in insufficient feature modeling. In this study, we propose a novel multimodal hybrid tracker (MMHT) that utilizes frame-event-based data for reliable single object tracking. The MMHT model employs a hybrid backbone consisting of an artificial neural network (ANN) and a spiking neural network (SNN) to extract dominant features from different visual modalities and then uses a unified encoder to align the features across different domains. Moreover, we propose an enhanced transformer-based module to fuse multimodal features using attention mechanisms. With these methods, the MMHT model can effectively construct a multiscale and multidimensional visual feature space and achieve discriminative feature modeling. Extensive experiments demonstrate that the MMHT model exhibits competitive performance in comparison with that of other state-of-the-art methods. Overall, our results highlight the effectiveness of the MMHT model in terms of addressing the challenges faced in visual object tracking tasks.
Collapse
Affiliation(s)
- Hongze Sun
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Rui Liu
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Wuque Cai
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Jun Wang
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Yue Wang
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China
| | - Huajin Tang
- College of Computer Science and Technology, Zhejiang University, Hangzhou 310027, China
| | - Yan Cui
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China; Sichuan Academy of Medical Sciences and Sichuan Provincial People's Hospital, Chengdu 611731, China
| | - Dezhong Yao
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China; Research Unit of NeuroInformation (2019RU035), Chinese Academy of Medical Sciences, Chengdu 611731, China.
| | - Daqing Guo
- Clinical Hospital of Chengdu Brain Science Institute, MOE Key Lab for NeuroInformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 611731, China.
| |
Collapse
|
3
|
Chen Y, Feng R, Xiong Z, Xiao J, Liu JK. High-performance deep spiking neural networks via at-most-two-spike exponential coding. Neural Netw 2024; 176:106346. [PMID: 38713970 DOI: 10.1016/j.neunet.2024.106346] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Revised: 04/18/2024] [Accepted: 04/25/2024] [Indexed: 05/09/2024]
Abstract
Spiking neural networks (SNNs) provide necessary models and algorithms for neuromorphic computing. A popular way of building high-performance deep SNNs is to convert ANNs to SNNs, taking advantage of advanced and well-trained ANNs. Here we propose an ANN to SNN conversion methodology that uses a time-based coding scheme, named At-most-two-spike Exponential Coding (AEC), and a corresponding AEC spiking neuron model for ANN-SNN conversion. AEC neurons employ quantization-compensating spikes to improve coding accuracy and capacity, with each neuron generating up to two spikes within the time window. Two exponential decay functions with tunable parameters are proposed to represent the dynamic encoding thresholds, based on which pixel intensities are encoded into spike times and spike times are decoded into pixel intensities. The hyper-parameters of AEC neurons are fine-tuned based on the loss function of SNN-decoded values and ANN-activation values. In addition, we design two regularization terms for the number of spikes, providing the possibility to achieve the best trade-off between accuracy, latency and power consumption. The experimental results show that, compared to other similar methods, the proposed scheme not only obtains deep SNNs with higher accuracy, but also has more significant advantages in terms of energy efficiency and inference latency. More details can be found at https://github.com/RPDS2020/AEC.git.
Collapse
Affiliation(s)
- Yunhua Chen
- School of Computer Science and Technology, Guangdong University of Technology, China.
| | - Ren Feng
- School of Computer Science and Technology, Guangdong University of Technology, China.
| | - Zhimin Xiong
- School of Computer Science and Technology, Guangdong University of Technology, China.
| | - Jinsheng Xiao
- School of Electronic Information, Wuhan University, China.
| | - Jian K Liu
- School of Computer Science, University of Birmingham, UK.
| |
Collapse
|
4
|
Wang Z, Ghaleb FA, Zainal A, Siraj MM, Lu X. An efficient intrusion detection model based on convolutional spiking neural network. Sci Rep 2024; 14:7054. [PMID: 38528084 DOI: 10.1038/s41598-024-57691-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Accepted: 03/20/2024] [Indexed: 03/27/2024] Open
Abstract
Many intrusion detection techniques have been developed to ensure that the target system can function properly under the established rules. With the booming Internet of Things (IoT) applications, the resource-constrained nature of its devices makes it urgent to explore lightweight and high-performance intrusion detection models. Recent years have seen a particularly active application of deep learning (DL) techniques. The spiking neural network (SNN), a type of artificial intelligence that is associated with sparse computations and inherent temporal dynamics, has been viewed as a potential candidate for the next generation of DL. It should be noted, however, that current research into SNNs has largely focused on scenarios where limited computational resources and insufficient power sources are not considered. Consequently, even state-of-the-art SNN solutions tend to be inefficient. In this paper, a lightweight and effective detection model is proposed. With the help of rational algorithm design, the model integrates the advantages of SNNs as well as convolutional neural networks (CNNs). In addition to reducing resource usage, it maintains a high level of classification accuracy. The proposed model was evaluated against some current state-of-the-art models using a comprehensive set of metrics. Based on the experimental results, the model demonstrated improved adaptability to environments with limited computational resources and energy sources.
Collapse
Affiliation(s)
- Zhen Wang
- Faculty of Computing, Universiti Teknologi Malaysia, Johor Bahru, 81310, Johor, Malaysia
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325035, Zhejiang, China
| | - Fuad A Ghaleb
- College of Computing and Digital Technology, Birmingham City University, Birmingham, B47XG, United Kingdom
| | - Anazida Zainal
- Faculty of Computing, Universiti Teknologi Malaysia, Johor Bahru, 81310, Johor, Malaysia
| | - Maheyzah Md Siraj
- Faculty of Computing, Universiti Teknologi Malaysia, Johor Bahru, 81310, Johor, Malaysia
| | - Xing Lu
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325035, Zhejiang, China.
| |
Collapse
|
5
|
Bacho F, Chu D. Exploring Trade-Offs in Spiking Neural Networks. Neural Comput 2023; 35:1627-1656. [PMID: 37523463 DOI: 10.1162/neco_a_01609] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Accepted: 06/03/2023] [Indexed: 08/02/2023]
Abstract
Spiking neural networks (SNNs) have emerged as a promising alternative to traditional deep neural networks for low-power computing. However, the effectiveness of SNNs is not solely determined by their performance but also by their energy consumption, prediction speed, and robustness to noise. The recent method Fast & Deep, along with others, achieves fast and energy-efficient computation by constraining neurons to fire at most once. Known as time-to-first-spike (TTFS), this constraint, however, restricts the capabilities of SNNs in many aspects. In this work, we explore the relationships of performance, energy consumption, speed, and stability when using this constraint. More precisely, we highlight the existence of trade-offs where performance and robustness are gained at the cost of sparsity and prediction latency. To improve these trade-offs, we propose a relaxed version of Fast & Deep that allows for multiple spikes per neuron. Our experiments show that relaxing the spike constraint provides higher performance while also benefiting from faster convergence, similar sparsity, comparable prediction latency, and better robustness to noise compared to TTFS SNNs. By highlighting the limitations of TTFS and demonstrating the advantages of unconstrained SNNs, we provide valuable insight for the development of effective learning strategies for neuromorphic computing.
Collapse
Affiliation(s)
- Florian Bacho
- CEMS, School of Computing, University of Kent, Canterbury CT2 7NF, U.K.
| | - Dominique Chu
- CEMS, School of Computing, University of Kent, Canterbury CT2 7NF, U.K.
| |
Collapse
|
6
|
Makarov VA, Lobov SA, Shchanikov S, Mikhaylov A, Kazantsev VB. Toward Reflective Spiking Neural Networks Exploiting Memristive Devices. Front Comput Neurosci 2022; 16:859874. [PMID: 35782090 PMCID: PMC9243340 DOI: 10.3389/fncom.2022.859874] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Accepted: 05/10/2022] [Indexed: 11/29/2022] Open
Abstract
The design of modern convolutional artificial neural networks (ANNs) composed of formal neurons copies the architecture of the visual cortex. Signals proceed through a hierarchy, where receptive fields become increasingly more complex and coding sparse. Nowadays, ANNs outperform humans in controlled pattern recognition tasks yet remain far behind in cognition. In part, it happens due to limited knowledge about the higher echelons of the brain hierarchy, where neurons actively generate predictions about what will happen next, i.e., the information processing jumps from reflex to reflection. In this study, we forecast that spiking neural networks (SNNs) can achieve the next qualitative leap. Reflective SNNs may take advantage of their intrinsic dynamics and mimic complex, not reflex-based, brain actions. They also enable a significant reduction in energy consumption. However, the training of SNNs is a challenging problem, strongly limiting their deployment. We then briefly overview new insights provided by the concept of a high-dimensional brain, which has been put forward to explain the potential power of single neurons in higher brain stations and deep SNN layers. Finally, we discuss the prospect of implementing neural networks in memristive systems. Such systems can densely pack on a chip 2D or 3D arrays of plastic synaptic contacts directly processing analog information. Thus, memristive devices are a good candidate for implementing in-memory and in-sensor computing. Then, memristive SNNs can diverge from the development of ANNs and build their niche, cognitive, or reflective computations.
Collapse
Affiliation(s)
- Valeri A. Makarov
- Instituto de Matemática Interdisciplinar, Universidad Complutense de Madrid, Madrid, Spain
- Department of Neurotechnologies, Research Institute of Physics and Technology, Laboratory of Stochastic Multistable Systems, Lobachevsky State University of Nizhny Novgorod, Nizhny Novgorod, Russia
| | - Sergey A. Lobov
- Department of Neurotechnologies, Research Institute of Physics and Technology, Laboratory of Stochastic Multistable Systems, Lobachevsky State University of Nizhny Novgorod, Nizhny Novgorod, Russia
- Neuroscience and Cognitive Technology Laboratory, Center for Technologies in Robotics and Mechatronics Components, Innopolis University, Innopolis, Russia
- Center For Neurotechnology and Machine Learning, Immanuel Kant Baltic Federal University, Kaliningrad, Russia
| | - Sergey Shchanikov
- Department of Neurotechnologies, Research Institute of Physics and Technology, Laboratory of Stochastic Multistable Systems, Lobachevsky State University of Nizhny Novgorod, Nizhny Novgorod, Russia
- Department of Information Technologies, Vladimir State University, Vladimir, Russia
| | - Alexey Mikhaylov
- Department of Neurotechnologies, Research Institute of Physics and Technology, Laboratory of Stochastic Multistable Systems, Lobachevsky State University of Nizhny Novgorod, Nizhny Novgorod, Russia
| | - Viktor B. Kazantsev
- Department of Neurotechnologies, Research Institute of Physics and Technology, Laboratory of Stochastic Multistable Systems, Lobachevsky State University of Nizhny Novgorod, Nizhny Novgorod, Russia
- Neuroscience and Cognitive Technology Laboratory, Center for Technologies in Robotics and Mechatronics Components, Innopolis University, Innopolis, Russia
- Center For Neurotechnology and Machine Learning, Immanuel Kant Baltic Federal University, Kaliningrad, Russia
| |
Collapse
|
7
|
Optimizing the Neural Structure and Hyperparameters of Liquid State Machines Based on Evolutionary Membrane Algorithm. MATHEMATICS 2022. [DOI: 10.3390/math10111844] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/10/2022]
Abstract
As one of the important artificial intelligence fields, brain-like computing attempts to give machines a higher intelligence level by studying and simulating the cognitive principles of the human brain. A spiking neural network (SNN) is one of the research directions of brain-like computing, characterized by better biogenesis and stronger computing power than the traditional neural network. A liquid state machine (LSM) is a neural computing model with a recurrent network structure based on SNN. In this paper, a learning algorithm based on an evolutionary membrane algorithm is proposed to optimize the neural structure and hyperparameters of an LSM. First, the object of the proposed algorithm is designed according to the neural structure and hyperparameters of the LSM. Second, the reaction rules of the proposed algorithm are employed to discover the best neural structure and hyperparameters of the LSM. Third, the membrane structure is that the skin membrane contains several elementary membranes to speed up the search of the proposed algorithm. In the simulation experiment, effectiveness verification is carried out on the MNIST and KTH datasets. In terms of the MNIST datasets, the best test results of the proposed algorithm with 500, 1000 and 2000 spiking neurons are 86.8%, 90.6% and 90.8%, respectively. The best test results of the proposed algorithm on KTH with 500, 1000 and 2000 spiking neurons are 82.9%, 85.3% and 86.3%, respectively. The simulation results show that the proposed algorithm has a more competitive advantage than other experimental algorithms.
Collapse
|