1
|
Kumari V, Kumar N, Kumar K S, Kumar A, Skandha SS, Saxena S, Khanna NN, Laird JR, Singh N, Fouda MM, Saba L, Singh R, Suri JS. Deep Learning Paradigm and Its Bias for Coronary Artery Wall Segmentation in Intravascular Ultrasound Scans: A Closer Look. J Cardiovasc Dev Dis 2023; 10:485. [PMID: 38132653 PMCID: PMC10743870 DOI: 10.3390/jcdd10120485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 10/15/2023] [Accepted: 11/07/2023] [Indexed: 12/23/2023] Open
Abstract
BACKGROUND AND MOTIVATION Coronary artery disease (CAD) has the highest mortality rate; therefore, its diagnosis is vital. Intravascular ultrasound (IVUS) is a high-resolution imaging solution that can image coronary arteries, but the diagnosis software via wall segmentation and quantification has been evolving. In this study, a deep learning (DL) paradigm was explored along with its bias. METHODS Using a PRISMA model, 145 best UNet-based and non-UNet-based methods for wall segmentation were selected and analyzed for their characteristics and scientific and clinical validation. This study computed the coronary wall thickness by estimating the inner and outer borders of the coronary artery IVUS cross-sectional scans. Further, the review explored the bias in the DL system for the first time when it comes to wall segmentation in IVUS scans. Three bias methods, namely (i) ranking, (ii) radial, and (iii) regional area, were applied and compared using a Venn diagram. Finally, the study presented explainable AI (XAI) paradigms in the DL framework. FINDINGS AND CONCLUSIONS UNet provides a powerful paradigm for the segmentation of coronary walls in IVUS scans due to its ability to extract automated features at different scales in encoders, reconstruct the segmented image using decoders, and embed the variants in skip connections. Most of the research was hampered by a lack of motivation for XAI and pruned AI (PAI) models. None of the UNet models met the criteria for bias-free design. For clinical assessment and settings, it is necessary to move from a paper-to-practice approach.
Collapse
Affiliation(s)
- Vandana Kumari
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (S.K.K.)
| | - Naresh Kumar
- Department of Applied Computational Science and Engineering, G L Bajaj Institute of Technology and Management, Greater Noida 201310, India
| | - Sampath Kumar K
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (S.K.K.)
| | - Ashish Kumar
- School of CSET, Bennett University, Greater Noida 201310, India;
| | - Sanagala S. Skandha
- Department of CSE, CMR College of Engineering and Technology, Hyderabad 501401, India;
| | - Sanjay Saxena
- Department of Computer Science and Engineering, IIT Bhubaneswar, Bhubaneswar 751003, India;
| | - Narendra N. Khanna
- Department of Cardiology, Indraprastha APOLLO Hospitals, New Delhi 110076, India;
| | - John R. Laird
- Heart and Vascular Institute, Adventist Health St. Helena, St Helena, CA 94574, USA;
| | - Narpinder Singh
- Department of Food Science and Technology, Graphic Era, Deemed to be University, Dehradun 248002, India;
| | - Mostafa M. Fouda
- Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA;
| | - Luca Saba
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), 09100 Cagliari, Italy;
| | - Rajesh Singh
- Department of Research and Innovation, Uttaranchal Institute of Technology, Uttaranchal University, Dehradun 248007, India;
| | - Jasjit S. Suri
- Stroke Diagnostics and Monitoring Division, AtheroPoint™, Roseville, CA 95661, USA
- Department of Computer Science & Engineering, Graphic Era, Deemed to be University, Dehradun 248002, India
- Monitoring and Diagnosis Division, AtheroPoint™, Roseville, CA 95661, USA
| |
Collapse
|
2
|
Gao S, Zhou M, Wang Z, Sugiyama D, Cheng J, Wang J, Todo Y. Fully Complex-Valued Dendritic Neuron Model. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:2105-2118. [PMID: 34487498 DOI: 10.1109/tnnls.2021.3105901] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
A single dendritic neuron model (DNM) that owns the nonlinear information processing ability of dendrites has been widely used for classification and prediction. Complex-valued neural networks that consist of a number of multiple/deep-layer McCulloch-Pitts neurons have achieved great successes so far since neural computing was utilized for signal processing. Yet no complex value representations appear in single neuron architectures. In this article, we first extend DNM from a real-value domain to a complex-valued one. Performance of complex-valued DNM (CDNM) is evaluated through a complex XOR problem, a non-minimum phase equalization problem, and a real-world wind prediction task. Also, a comparative analysis on a set of elementary transcendental functions as an activation function is implemented and preparatory experiments are carried out for determining hyperparameters. The experimental results indicate that the proposed CDNM significantly outperforms real-valued DNM, complex-valued multi-layer perceptron, and other complex-valued neuron models.
Collapse
|
3
|
Li C, Liu S, Wang Z, Yuan G. Classifying epileptic phase-amplitude coupling in SEEG using complex-valued convolutional neural network. Front Physiol 2023; 13:1085530. [PMID: 36685186 PMCID: PMC9849379 DOI: 10.3389/fphys.2022.1085530] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Accepted: 12/20/2022] [Indexed: 01/06/2023] Open
Abstract
EEG phase-amplitude coupling (PAC), the amplitude of high-frequency oscillations modulated by the phase of low-frequency oscillations (LFOs), is a useful biomarker to localize epileptogenic tissue. It is commonly represented in a comodulogram of coupling strength but without coupled phase information. The phase-amplitude coupling is also found in the normal brain, and it is difficult to discriminate pathological phase-amplitude couplings from normal ones. This study proposes a novel approach based on complex-valued phase-amplitude coupling (CV-PAC) for classifying epileptic phase-amplitude coupling. The CV-PAC combines both the coupling strengths and the coupled phases of low-frequency oscillations. The complex-valued convolutional neural network (CV-CNN) is then used to classify epileptic CV-PAC. Stereo-electroencephalography (SEEG) recordings from nine intractable epilepsy patients were analyzed. The leave-one-out cross-validation is performed, and the area-under-curve (AUC) value is used as the indicator of the performance of different measures. Our result shows that the area-under-curve value is .92 for classifying epileptic CV-PAC using CV-CNN. The area-under-curve value decreases to .89, .80, and .88 while using traditional convolutional neural networks, support vector machine, and random forest, respectively. The phases of delta (1-4 Hz) and alpha (8-10 Hz) bands are different between epileptic and normal CV-PAC. The phase information of CV-PAC is important for improving classification performance. The proposed approach of CV-PAC/CV-CNN promises to identify more accurate epileptic brain activities for potential surgical intervention.
Collapse
Affiliation(s)
- Chunsheng Li
- Department of Biomedical Engineering, School of Electrical Engineering, Shenyang University of Technology, Shenyang, China,*Correspondence: Chunsheng Li,
| | - Shiyue Liu
- Department of Biomedical Engineering, School of Electrical Engineering, Shenyang University of Technology, Shenyang, China
| | - Zeyu Wang
- Department of Biomedical Engineering, School of Electrical Engineering, Shenyang University of Technology, Shenyang, China,Department of Electrical Engineering and Information Systems, University of Pannonia, Veszprem, Hungary
| | - Guanqian Yuan
- Department of Neurosurgery, General Hospital of Northern Theater Command, Shenyang, China
| |
Collapse
|
4
|
Nguyen K, Fookes C, Sridharan S, Ross A. Complex-Valued Iris Recognition Network. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:182-196. [PMID: 35201979 DOI: 10.1109/tpami.2022.3152857] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
In this work, we design a fully complex-valued neural network for the task of iris recognition. Unlike the problem of general object recognition, where real-valued neural networks can be used to extract pertinent features, iris recognition depends on the extraction of both phase and magnitude information from the input iris texture in order to better represent its biometric content. This necessitates the extraction and processing of phase information that cannot be effectively handled by a real-valued neural network. In this regard, we design a fully complex-valued neural network that can better capture the multi-scale, multi-resolution, and multi-orientation phase and amplitude features of the iris texture. We show a strong correspondence of the proposed complex-valued iris recognition network with Gabor wavelets that are used to generate the classical IrisCode; however, the proposed method enables a new capability of automatic complex-valued feature learning that is tailored for iris recognition. We conduct experiments on three benchmark datasets - ND-CrossSensor-2013, CASIA-Iris-Thousand and UBIRIS.v2 - and show the benefit of the proposed network for the task of iris recognition. We exploit visualization schemes to convey how the complex-valued network, when compared to standard real-valued networks, extracts fundamentally different features from the iris texture.
Collapse
|
5
|
Quan Y, Lin P, Xu Y, Nan Y, Ji H. Nonblind Image Deblurring via Deep Learning in Complex Field. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:5387-5400. [PMID: 33852398 DOI: 10.1109/tnnls.2021.3070596] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Nonblind image deblurring is about recovering the latent clear image from a blurry one generated by a known blur kernel, which is an often-seen yet challenging inverse problem in imaging. Its key is how to robustly suppress noise magnification during the inversion process. Recent approaches made a breakthrough by exploiting convolutional neural network (CNN)-based denoising priors in the image domain or the gradient domain, which allows using a CNN for noise suppression. The performance of these approaches is highly dependent on the effectiveness of the denoising CNN in removing magnified noise whose distribution is unknown and varies at different iterations of the deblurring process for different images. In this article, we introduce a CNN-based image prior defined in the Gabor domain. The prior not only utilizes the optimal space-frequency resolution and strong orientation selectivity of the Gabor transform but also enables using complex-valued (CV) representations in intermediate processing for better denoising. A CV CNN is developed to exploit the benefits of the CV representations, with better generalization to handle unknown noises over the real-valued ones. Combining our Gabor-domain CV CNN-based prior with an unrolling scheme, we propose a deep-learning-based approach to nonblind image deblurring. Extensive experiments have demonstrated the superior performance of the proposed approach over the state-of-the-art ones.
Collapse
|
6
|
Image classification based on quaternion-valued capsule network. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03849-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
7
|
Zhang SQ, Gao W, Zhou ZH. Towards understanding theoretical advantages of complex-reaction networks. Neural Netw 2022; 151:80-93. [DOI: 10.1016/j.neunet.2022.03.024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Revised: 01/21/2022] [Accepted: 03/22/2022] [Indexed: 11/17/2022]
|
8
|
A study on generic object detection with emphasis on future research directions. JOURNAL OF KING SAUD UNIVERSITY - COMPUTER AND INFORMATION SCIENCES 2021. [DOI: 10.1016/j.jksuci.2021.08.001] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
9
|
Gutiérrez J, Rodríguez V, Martin S. Comprehensive Review of Vision-Based Fall Detection Systems. SENSORS (BASEL, SWITZERLAND) 2021; 21:947. [PMID: 33535373 PMCID: PMC7866979 DOI: 10.3390/s21030947] [Citation(s) in RCA: 23] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/18/2020] [Revised: 01/19/2021] [Accepted: 01/25/2021] [Indexed: 11/16/2022]
Abstract
Vision-based fall detection systems have experienced fast development over the last years. To determine the course of its evolution and help new researchers, the main audience of this paper, a comprehensive revision of all published articles in the main scientific databases regarding this area during the last five years has been made. After a selection process, detailed in the Materials and Methods Section, eighty-one systems were thoroughly reviewed. Their characterization and classification techniques were analyzed and categorized. Their performance data were also studied, and comparisons were made to determine which classifying methods best work in this field. The evolution of artificial vision technology, very positively influenced by the incorporation of artificial neural networks, has allowed fall characterization to become more resistant to noise resultant from illumination phenomena or occlusion. The classification has also taken advantage of these networks, and the field starts using robots to make these systems mobile. However, datasets used to train them lack real-world data, raising doubts about their performances facing real elderly falls. In addition, there is no evidence of strong connections between the elderly and the communities of researchers.
Collapse
Affiliation(s)
- Jesús Gutiérrez
- Universidad Nacional de Educación a Distancia, Juan Rosal 12, 28040 Madrid, Spain;
| | - Víctor Rodríguez
- EduQTech, E.U. Politécnica, Maria Lluna 3, 50018 Zaragoza, Spain;
| | - Sergio Martin
- Universidad Nacional de Educación a Distancia, Juan Rosal 12, 28040 Madrid, Spain;
| |
Collapse
|
10
|
Caramazza P, Moran O, Murray-Smith R, Faccio D. Transmission of natural scene images through a multimode fibre. Nat Commun 2019; 10:2029. [PMID: 31048712 PMCID: PMC6497636 DOI: 10.1038/s41467-019-10057-8] [Citation(s) in RCA: 48] [Impact Index Per Article: 9.6] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2018] [Accepted: 04/11/2019] [Indexed: 12/20/2022] Open
Abstract
The optical transport of images through a multimode fibre remains an outstanding challenge with applications ranging from optical communications to neuro-imaging. State of the art approaches either involve measurement and control of the full complex field transmitted through the fibre or, more recently, training of artificial neural networks that however, are typically limited to image classes belong to the same class as the training data set. Here we implement a method that statistically reconstructs the inverse transformation matrix for the fibre. We demonstrate imaging at high frame rates, high resolutions and in full colour of natural scenes, thus demonstrating general-purpose imaging capability. Real-time imaging over long fibre lengths opens alternative routes to exploitation for example for secure communication systems, novel remote imaging devices, quantum state control processing and endoscopy.
Collapse
Affiliation(s)
| | - Oisín Moran
- School of Computing Science, University of Glasgow, Glasgow, G12 8QQ, UK
| | | | - Daniele Faccio
- School of Physics and Astronomy, University of Glasgow, Glasgow, G12 8QQ, UK.
| |
Collapse
|
11
|
Lundervold AS, Lundervold A. An overview of deep learning in medical imaging focusing on MRI. Z Med Phys 2018; 29:102-127. [PMID: 30553609 DOI: 10.1016/j.zemedi.2018.11.002] [Citation(s) in RCA: 705] [Impact Index Per Article: 117.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2018] [Revised: 11/19/2018] [Accepted: 11/21/2018] [Indexed: 02/06/2023]
Abstract
What has happened in machine learning lately, and what does it mean for the future of medical image analysis? Machine learning has witnessed a tremendous amount of attention over the last few years. The current boom started around 2009 when so-called deep artificial neural networks began outperforming other established models on a number of important benchmarks. Deep neural networks are now the state-of-the-art machine learning models across a variety of areas, from image analysis to natural language processing, and widely deployed in academia and industry. These developments have a huge potential for medical imaging technology, medical data analysis, medical diagnostics and healthcare in general, slowly being realized. We provide a short overview of recent advances and some associated challenges in machine learning applied to medical image processing and image analysis. As this has become a very broad and fast expanding field we will not survey the entire landscape of applications, but put particular focus on deep learning in MRI. Our aim is threefold: (i) give a brief introduction to deep learning with pointers to core references; (ii) indicate how deep learning has been applied to the entire MRI processing chain, from acquisition to image retrieval, from segmentation to disease prediction; (iii) provide a starting point for people interested in experimenting and perhaps contributing to the field of deep learning for medical imaging by pointing out good educational resources, state-of-the-art open-source code, and interesting sources of data and problems related medical imaging.
Collapse
Affiliation(s)
- Alexander Selvikvåg Lundervold
- Mohn Medical Imaging and Visualization Centre (MMIV), Haukeland University Hospital, Norway; Department of Computing, Mathematics and Physics, Western Norway University of Applied Sciences, Norway.
| | - Arvid Lundervold
- Mohn Medical Imaging and Visualization Centre (MMIV), Haukeland University Hospital, Norway; Neuroinformatics and Image Analysis Laboratory, Department of Biomedicine, University of Bergen, Norway; Department of Health and Functioning, Western Norway University of Applied Sciences, Norway.
| |
Collapse
|
12
|
Zhang J, Wu Y. Complex-valued unsupervised convolutional neural networks for sleep stage classification. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 164:181-191. [PMID: 30195426 DOI: 10.1016/j.cmpb.2018.07.015] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/26/2018] [Revised: 07/05/2018] [Accepted: 07/25/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE Despite numerous deep learning methods being developed for automatic sleep stage classification, almost all the models need labeled data. However, obtaining labeled data is a subjective process. Therefore, the labels will be different between two experts. At the same time, obtaining labeled data also is a time-consuming task. Even an experienced expert requires hours to annotate the sleep stage patterns. More important, as the development of wearable sleep devices, it is very difficult to obtain labeled sleep data. Therefore, unsupervised training algorithm is very important for sleep stage classification. Hence, a new sleep stage classification method named complex-valued unsupervised convolutional neural networks (CUCNN) is proposed in this study. METHODS The CUCNN operates with complex-valued inputs, outputs, and weights, and its training strategy is greedy layer-wise training. It is composed of three phases: phase encoder, unsupervised training and complex-valued classification. Phase encoder is used to translate real-valued inputs into complex numbers. In the unsupervised training phase, the complex-valued K-means is used to learn filters which will be used in the convolution. RESULTS The classification performances of handcrafted features are compared with those of learned features via CUCNN. The total accuracy (TAC) and kappa coefficient of the sleep stage from UCD dataset are 87% and 0.8, respectively. Moreover, the comparison experiments indicate that the TACs of the CUCNN from UCD and MIT-BIH datasets outperform these of unsupervised convolutional neural networks (UCNN) by 12.9% and 13%, respectively. Additionally, the convergence of CUCNN is much faster than that of UCNN in most cases. CONCLUSIONS The proposed method is fully automated and can learn features in an unsupervised fashion. Results show that unsupervised training and automatic feature extraction on sleep data are possible, which are very important for home sleep monitoring.
Collapse
Affiliation(s)
- Junming Zhang
- College of Electronics & Information Engineering, Tongji University, Shanghai 201804, China
| | - Yan Wu
- College of Electronics & Information Engineering, Tongji University, Shanghai 201804, China.
| |
Collapse
|
13
|
Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E. Deep Learning for Computer Vision: A Brief Review. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2018; 2018:7068349. [PMID: 29487619 PMCID: PMC5816885 DOI: 10.1155/2018/7068349] [Citation(s) in RCA: 591] [Impact Index Per Article: 98.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/17/2017] [Accepted: 11/27/2017] [Indexed: 12/13/2022]
Abstract
Over the last years deep learning methods have been shown to outperform previous state-of-the-art machine learning techniques in several fields, with computer vision being one of the most prominent cases. This review paper provides a brief overview of some of the most significant deep learning schemes used in computer vision problems, that is, Convolutional Neural Networks, Deep Boltzmann Machines and Deep Belief Networks, and Stacked Denoising Autoencoders. A brief account of their history, structure, advantages, and limitations is given, followed by a description of their applications in various computer vision tasks, such as object detection, face recognition, action and activity recognition, and human pose estimation. Finally, a brief overview is given of future directions in designing deep learning schemes for computer vision problems and the challenges involved therein.
Collapse
Affiliation(s)
- Athanasios Voulodimos
- Department of Informatics, Technological Educational Institute of Athens, 12210 Athens, Greece
- National Technical University of Athens, 15780 Athens, Greece
| | | | | | | |
Collapse
|
14
|
|