1
|
Li Y, El Habib Daho M, Conze PH, Zeghlache R, Le Boité H, Tadayoni R, Cochener B, Lamard M, Quellec G. A review of deep learning-based information fusion techniques for multimodal medical image classification. Comput Biol Med 2024; 177:108635. [PMID: 38796881 DOI: 10.1016/j.compbiomed.2024.108635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 03/18/2024] [Accepted: 05/18/2024] [Indexed: 05/29/2024]
Abstract
Multimodal medical imaging plays a pivotal role in clinical diagnosis and research, as it combines information from various imaging modalities to provide a more comprehensive understanding of the underlying pathology. Recently, deep learning-based multimodal fusion techniques have emerged as powerful tools for improving medical image classification. This review offers a thorough analysis of the developments in deep learning-based multimodal fusion for medical classification tasks. We explore the complementary relationships among prevalent clinical modalities and outline three main fusion schemes for multimodal classification networks: input fusion, intermediate fusion (encompassing single-level fusion, hierarchical fusion, and attention-based fusion), and output fusion. By evaluating the performance of these fusion techniques, we provide insight into the suitability of different network architectures for various multimodal fusion scenarios and application domains. Furthermore, we delve into challenges related to network architecture selection, handling incomplete multimodal data management, and the potential limitations of multimodal fusion. Finally, we spotlight the promising future of Transformer-based multimodal fusion techniques and give recommendations for future research in this rapidly evolving field.
Collapse
Affiliation(s)
- Yihao Li
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Mostafa El Habib Daho
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France.
| | | | - Rachid Zeghlache
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Hugo Le Boité
- Sorbonne University, Paris, France; Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France
| | - Ramin Tadayoni
- Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France; Paris Cité University, Paris, France
| | - Béatrice Cochener
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France; Ophthalmology Department, CHRU Brest, Brest, France
| | - Mathieu Lamard
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | | |
Collapse
|
2
|
Ma H, Wu Y, Tang Y, Chen R, Xu T, Zhang W. Parallel Dual-Branch Fusion Network for Epileptic Seizure Prediction. Comput Biol Med 2024; 176:108565. [PMID: 38744007 DOI: 10.1016/j.compbiomed.2024.108565] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 04/09/2024] [Accepted: 05/05/2024] [Indexed: 05/16/2024]
Abstract
Epilepsy is a prevalent chronic disorder of the central nervous system. The timely and accurate seizure prediction using the scalp Electroencephalography (EEG) signal can make patients adopt reasonable preventive measures before seizures occur and thus reduce harm to patients. In recent years, deep learning-based methods have made significant progress in solving the problem of epileptic seizure prediction. However, most current methods mainly focus on modeling short- or long-term dependence in EEG, while neglecting to consider both. In this study, we propose a Parallel Dual-Branch Fusion Network (PDBFusNet) which aims to combine the complementary advantages of Convolutional Neural Network (CNN) and Transformer. Specifically, the features of the EEG signal are first extracted using Mel Frequency Cepstral Coefficients (MFCC). Then, the extracted features are delivered into the parallel dual-branches to simultaneously capture the short- and long-term dependencies of EEG signal. Further, regarding the Transformer branch, a novel feature fusion module is developed to enhance the ability of utilizing time, frequency, and channel information. To evaluate our proposal, we perform sufficient experiments on the public epileptic EEG dataset CHB-MIT, where the accuracy, sensitivity, specificity and precision are 95.76%, 95.81%, 95.71% and 95.71%, respectively. PDBFusNet shows superior performance compared to state-of-the-art competitors, which confirms the effectiveness of our proposal.
Collapse
Affiliation(s)
- Hongcheng Ma
- School of Information and Communication Engineering, Hainan University, Haikou, China; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China.
| | - Yajing Wu
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China.
| | - Yongqiang Tang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China.
| | - Rui Chen
- School of Information and Communication Engineering, Hainan University, Haikou, China; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China.
| | - Tao Xu
- Shanxi Key Laboratory of Big Data Analysis and Parallel Computing, Taiyuan University of Science and Technology, Taiyuan, China.
| | - Wensheng Zhang
- School of Information and Communication Engineering, Hainan University, Haikou, China; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou, China.
| |
Collapse
|
3
|
Li N, Xiao J, Mao N, Cheng D, Chen X, Zhao F, Shi Z. Joint learning of multi-level dynamic brain networks for autism spectrum disorder diagnosis. Comput Biol Med 2024; 171:108054. [PMID: 38350396 DOI: 10.1016/j.compbiomed.2024.108054] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 01/08/2024] [Accepted: 01/26/2024] [Indexed: 02/15/2024]
Abstract
Graph convolutional networks (GCNs), with their powerful ability to model non-Euclidean graph data, have shown advantages in learning representations of brain networks. However, considering the complexity, multilayeredness, and spatio-temporal dynamics of brain activities, we have identified two limitations in current GCN-based research on brain networks: 1) Most studies have focused on unidirectional information transmission across brain network levels, neglecting joint learning or bidirectional information exchange among networks. 2) Most of the existing models determine node neighborhoods by thresholding or simply binarizing the brain network, which leads to the loss of edge weight information and weakens the model's sensitivity to important information in the brain network. To address the above issues, we propose a multi-level dynamic brain network joint learning architecture based on GCN for autism spectrum disorder (ASD) diagnosis. Specifically, firstly, constructing different-level dynamic brain networks. Then, utilizing joint learning based on GCN for interactive information exchange among these multi-level brain networks. Finally, designing an edge self-attention mechanism to assign different edge weights to inter-node connections, which allows us to pick out the crucial features for ASD diagnosis. Our proposed method achieves an accuracy of 81.5 %. The results demonstrate that our method enables bidirectional transfer of high-order and low-order information, facilitating information complementarity between different levels of brain networks. Additionally, the use of edge weights enhances the representation capability of ASD-related features.
Collapse
Affiliation(s)
- Na Li
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China; Department of Computer Science and Engineering, Xi'an University of Technology, Xi'an, Shanxi, China
| | - Jinjie Xiao
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China
| | - Ning Mao
- Department of Radiology, Yantai Yuhuangding Hospital, Yantai, China
| | - Dapeng Cheng
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China
| | - Xiaobo Chen
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China
| | - Feng Zhao
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China.
| | - Zhenghao Shi
- Department of Computer Science and Engineering, Xi'an University of Technology, Xi'an, Shanxi, China.
| |
Collapse
|
4
|
Wen W, Zhang H, Wang Z, Gao X, Wu P, Lin J, Zeng N. Enhanced multi-label cardiology diagnosis with channel-wise recurrent fusion. Comput Biol Med 2024; 171:108210. [PMID: 38417383 DOI: 10.1016/j.compbiomed.2024.108210] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Revised: 02/08/2024] [Accepted: 02/25/2024] [Indexed: 03/01/2024]
Abstract
The timely detection of abnormal electrocardiogram (ECG) signals is vital for preventing heart disease. However, traditional automated cardiology diagnostic methods have the limitation of being unable to simultaneously identify multiple diseases in a segment of ECG signals, and do not consider the potential correlations between the 12-lead ECG signals. To address these issues, this paper presents a novel network architecture, denoted as Branched Convolution and Channel Fusion Network (BCCF-Net), designed for the multi-label diagnosis of ECG cardiology to achieve simultaneous identification of multiple diseases. Among them, the BCCF-Net incorporates the Channel-wise Recurrent Fusion (CRF) network, which is designed to enhance the ability to explore potential correlation information between 12 leads. Furthermore, the utilization of the squeeze and excitation (SE) attention mechanism maximizes the potential of the convolutional neural network (CNN). In order to efficiently capture complex patterns in space and time across various scales, the multi branch convolution (MBC) module has been developed. Through extensive experiments on two public datasets with seven subtasks, the efficacy and robustness of the proposed ECG multi-label classification framework have been comprehensively evaluated. The results demonstrate the superior performance of the BCCF-Net compared to other state-of-the-art algorithms. The developed framework holds practical application in clinical settings, allowing for the refined diagnosis of cardiac arrhythmias through ECG signal analysis.
Collapse
Affiliation(s)
- Weimin Wen
- School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361024, China
| | - Hongyi Zhang
- School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361024, China
| | - Zidong Wang
- Department of Computer Science, Brunel University London, Uxbridge UB8 3PH, UK.
| | - Xingen Gao
- School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361024, China
| | - Peishu Wu
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361005, China
| | - Juqiang Lin
- School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361024, China
| | - Nianyin Zeng
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361005, China.
| |
Collapse
|
5
|
Guo X, Wang Z, Wu P, Li Y, Alsaadi FE, Zeng N. ELTS-Net: An enhanced liver tumor segmentation network with augmented receptive field and global contextual information. Comput Biol Med 2024; 169:107879. [PMID: 38142549 DOI: 10.1016/j.compbiomed.2023.107879] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Revised: 11/30/2023] [Accepted: 12/18/2023] [Indexed: 12/26/2023]
Abstract
The liver is one of the organs with the highest incidence rate in the human body, and late-stage liver cancer is basically incurable. Therefore, early diagnosis and lesion location of liver cancer are of important clinical value. This study proposes an enhanced network architecture ELTS-Net based on the 3D U-Net model, to address the limitations of conventional image segmentation methods and the underutilization of image spatial features by the 2D U-Net network structure. ELTS-Net expands upon the original network by incorporating dilated convolutions to increase the receptive field of the convolutional kernel. Additionally, an attention residual module, comprising an attention mechanism and residual connections, replaces the original convolutional module, serving as the primary components of the encoder and decoder. This design enables the network to capture contextual information globally in both channel and spatial dimensions. Furthermore, deep supervision modules are integrated between different levels of the decoder network, providing additional feedback from deeper intermediate layers. This constrains the network weights to the target regions and optimizing segmentation results. Evaluation on the LiTS2017 dataset shows improvements in evaluation metrics for liver and tumor segmentation tasks compared to the baseline 3D U-Net model, achieving 95.2% liver segmentation accuracy and 71.9% tumor segmentation accuracy, with accuracy improvements of 0.9% and 3.1% respectively. The experimental results validate the superior segmentation performance of ELTS-Net compared to other comparison models, offering valuable guidance for clinical diagnosis and treatment.
Collapse
Affiliation(s)
- Xiaoyue Guo
- College of Engineering, Peking University, Beijing 100871, China; Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361005, China
| | - Zidong Wang
- Department of Computer Science, Brunel University London, Uxbridge UB8 3PH, UK.
| | - Peishu Wu
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361005, China
| | - Yurong Li
- College of Electrical Engineering and Automation, Fuzhou University, Fujian 350116, China; Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fujian 350116, China
| | - Fuad E Alsaadi
- Communication Systems and Networks Research Group, Department of Electrical and Computer Engineering, Faculty of Engineering, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| | - Nianyin Zeng
- Department of Instrumental and Electrical Engineering, Xiamen University, Fujian 361005, China.
| |
Collapse
|
6
|
Pitarch C, Ungan G, Julià-Sapé M, Vellido A. Advances in the Use of Deep Learning for the Analysis of Magnetic Resonance Image in Neuro-Oncology. Cancers (Basel) 2024; 16:300. [PMID: 38254790 PMCID: PMC10814384 DOI: 10.3390/cancers16020300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2023] [Revised: 12/28/2023] [Accepted: 01/08/2024] [Indexed: 01/24/2024] Open
Abstract
Machine Learning is entering a phase of maturity, but its medical applications still lag behind in terms of practical use. The field of oncological radiology (and neuro-oncology in particular) is at the forefront of these developments, now boosted by the success of Deep-Learning methods for the analysis of medical images. This paper reviews in detail some of the most recent advances in the use of Deep Learning in this field, from the broader topic of the development of Machine-Learning-based analytical pipelines to specific instantiations of the use of Deep Learning in neuro-oncology; the latter including its use in the groundbreaking field of ultra-low field magnetic resonance imaging.
Collapse
Affiliation(s)
- Carla Pitarch
- Department of Computer Science, Universitat Politècnica de Catalunya (UPC BarcelonaTech) and Intelligent Data Science and Artificial Intelligence (IDEAI-UPC) Research Center, 08034 Barcelona, Spain;
- Eurecat, Digital Health Unit, Technology Centre of Catalonia, 08005 Barcelona, Spain
| | - Gulnur Ungan
- Departament de Bioquímica i Biologia Molecular and Institut de Biotecnologia i Biomedicina (IBB), Universitat Autònoma de Barcelona (UAB), 08193 Barcelona, Spain; (G.U.); (M.J.-S.)
- Centro de Investigación Biomédica en Red (CIBER), 28029 Madrid, Spain
| | - Margarida Julià-Sapé
- Departament de Bioquímica i Biologia Molecular and Institut de Biotecnologia i Biomedicina (IBB), Universitat Autònoma de Barcelona (UAB), 08193 Barcelona, Spain; (G.U.); (M.J.-S.)
- Centro de Investigación Biomédica en Red (CIBER), 28029 Madrid, Spain
| | - Alfredo Vellido
- Department of Computer Science, Universitat Politècnica de Catalunya (UPC BarcelonaTech) and Intelligent Data Science and Artificial Intelligence (IDEAI-UPC) Research Center, 08034 Barcelona, Spain;
- Centro de Investigación Biomédica en Red (CIBER), 28029 Madrid, Spain
| |
Collapse
|
7
|
Chen Q, Wang L, Xing Z, Wang L, Hu X, Wang R, Zhu YM. Deep wavelet scattering orthogonal fusion network for glioma IDH mutation status prediction. Comput Biol Med 2023; 166:107493. [PMID: 37774558 DOI: 10.1016/j.compbiomed.2023.107493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Revised: 06/26/2023] [Accepted: 09/15/2023] [Indexed: 10/01/2023]
Abstract
Accurately predicting the isocitrate dehydrogenase (IDH) mutation status of gliomas is greatly significant for formulating appropriate treatment plans and evaluating the prognoses of gliomas. Although existing studies can accurately predict the IDH mutation status of gliomas based on multimodal magnetic resonance (MR) images and machine learning methods, most of these methods cannot fully explore multimodal information and effectively predict IDH status for datasets acquired from multiple centers. To address this issue, a novel wavelet scattering (WS)-based orthogonal fusion network (WSOFNet) was proposed in this work to predict the IDH mutation status of gliomas from multiple centers. First, transformation-invariant features were extracted from multimodal MR images with a WS network, and then the multimodal WS features were used instead of the original images as the inputs of WSOFNet and were fully fused through an adaptive multimodal feature fusion module (AMF2M) and an orthogonal projection module (OPM). Finally, the fused features were input into a fully connected classifier to predict IDH mutation status. In addition, to achieve improved prediction accuracy, four auxiliary losses were also used in the feature extraction modules. The comparison results showed that the prediction area under the curve (AUC) of WSOFNet on a single-center dataset was 0.9966 and that on a multicenter dataset was approximately 0.9655, which was at least 3.9% higher than that of state-of-the-art methods. Moreover, the ablation experimental results also proved that the adaptive multimodal feature fusion strategy based on orthogonal projection could effectively improve the prediction performance of the model, especially for an external validation dataset.
Collapse
Affiliation(s)
- Qijian Chen
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Lihui Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China.
| | - Zhiyang Xing
- Department of Radiology, International Exemplary Cooperation Base of Precision Imaging for Diagnosis and Treatment, NHC Key Laboratory of Pulmonary Immune-related Diseases, Guizhou Provincial People's Hospital, Guiyang, 550002, China
| | - Li Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Xubin Hu
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
| | - Rongpin Wang
- Department of Radiology, International Exemplary Cooperation Base of Precision Imaging for Diagnosis and Treatment, NHC Key Laboratory of Pulmonary Immune-related Diseases, Guizhou Provincial People's Hospital, Guiyang, 550002, China
| | - Yue-Min Zhu
- University Lyon, INSA Lyon, CNRS, Inserm, IRP Metislab CREATIS UMR5220, U1206, Lyon 69621, France
| |
Collapse
|
8
|
Matsumura U, Tsurusaki T, Ogusu R, Yamamoto S, Lee Y, Sunagawa S, Reid WD, Koseki H. The interrelationship between lower limb movement, muscle activity, and joint moment during half squat and gait. Heliyon 2023; 9:e21762. [PMID: 38028012 PMCID: PMC10651501 DOI: 10.1016/j.heliyon.2023.e21762] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Revised: 10/26/2023] [Accepted: 10/27/2023] [Indexed: 12/01/2023] Open
Abstract
Joint moment is the resultant force of limb movements. However, estimation methods for joint moments using surface electromyography frequently use joint angles instead of limb angles. The limb angle in joint moment estimation using electromyography could clarify the effects of muscle activity on the limbs: acceleration, deceleration, or stabilization. No study has quantified the comprehensive relationship between limb movement, muscle activity, and joint moment. This study aimed to determine the influencing factors for ankle-joint moment and knee-joint moment in the sagittal plane among muscle activities and parameters related to limb movements during half squat and gait. This study included 29 healthy adults (16 female participants, 21.1 ± 2.09 years). Using inertial measurement units, thigh, shank, and foot inclination angles and angular accelerations were calculated as the parameters of limb movements. Muscle activations of the biceps femoris long head, rectus femoris, gastrocnemius, and tibialis anterior were measured. Ankle joint moment and knee-joint moment were measured using a three-dimensional motion capture system and two force plates. Regression models showed high accuracy in measuring ankle-joint moment during a half squat and gait (R2f = 0.92, 0.97, respectively) and knee-joint moment during a half squat (R2f = 0.98), but not knee-joint moment during gait (R2f = 0.63). However, only a maximum of five parameters were selected from muscle activities and limb angular information. Tibialis anterior and gastrocnemius activity were the largest contributors to ankle-joint moment during a half squat and gait, respectively, while muscle activities were not directly reflected in the knee-joint moment during either movement. Consideration of the interrelationships among limb movement, muscle activity, and joint moment is required when adjusting joint movements according to the target and aim of the therapeutic interventions.
Collapse
Affiliation(s)
- Umi Matsumura
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
| | - Toshiya Tsurusaki
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
| | - Rena Ogusu
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
| | - Shimpei Yamamoto
- Department of Rehabilitation, Fukuoka University Chikushi Hospital, Fukuoka, Japan
| | - Yeonghee Lee
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
| | - Shinya Sunagawa
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
- Department of Rehabilitation, Wajinkai Hospital, Nagasaki, Japan
| | - W Darlene Reid
- Department of Physical Therapy, University of Toronto, Toronto, ON, Canada
- Interdepartmental Division of Critical Care Medicine, University of Toronto, Toronto, ON, Canada
- KITE, Toronto Rehabilitation Institute, University Health Network, Toronto, ON, Canada
| | - Hironobu Koseki
- Department of Health Sciences, Nagasaki University Graduate School of Biomedical Sciences, Nagasaki, Japan
| |
Collapse
|
9
|
Xiong B, Chen W, Niu Y, Gan Z, Mao G, Xu Y. A Global and Local Feature fused CNN architecture for the sEMG-based hand gesture recognition. Comput Biol Med 2023; 166:107497. [PMID: 37783073 DOI: 10.1016/j.compbiomed.2023.107497] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2023] [Revised: 07/22/2023] [Accepted: 09/15/2023] [Indexed: 10/04/2023]
Abstract
Deep learning methods have been widely used for the classification of hand gestures using sEMG signals. Existing deep learning architectures only captures local spatial information and has limitations in extracting global temporal dependency to enhance the model's performance. In this paper, we propose a Global and Local Feature fused CNN (GLF-CNN) model that extracts features both globally and locally from sEMG signals to enhance the performance of hand gestures classification. The model contains two independent branches extracting local and global features each and fuses them to learn more diversified features and effectively improve the stability of gesture recognition. Besides, it also exhibits lower computational cost compared to the present approaches. We conduct experiments on five benchmark databases, including the NinaPro DB4, NinaPro DB5, BioPatRec DB1-DB3, and the Mendeley Data. The proposed model achieved the highest average accuracy of 88.34% on these databases, with a 9.96% average accuracy improvement and a 50% reduction in variance compared to the models with the same number of parameters. Moreover, the classification accuracies for the BioPatRec DB1, BioPatRec DB3 and Mendeley Data are 91.4%, 91.0% and 88.6% respectively, corresponding to an improvement of 13.2%, 41.5% and 12.2% over the respective state-of-the-art models. The experimental results demonstrate that the proposed model effectively enhances robustness, with improved gesture recognition performance and generalization ability. It contributes a new way for prosthetic control and human-machine interaction.
Collapse
Affiliation(s)
- Baoping Xiong
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China
| | - Wensheng Chen
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China
| | - Yinxi Niu
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China
| | - Zhenhua Gan
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China
| | - Guojun Mao
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China
| | - Yong Xu
- Computer Science and Mathematics, Fujian University of Technology, Fujian 350116, China.
| |
Collapse
|
10
|
Hsieh C, Nobre IB, Sousa SC, Ouyang C, Brereton M, Nascimento JC, Jorge J, Moreira C. MDF-Net for abnormality detection by fusing X-rays with clinical data. Sci Rep 2023; 13:15873. [PMID: 37741833 PMCID: PMC10517966 DOI: 10.1038/s41598-023-41463-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2023] [Accepted: 08/27/2023] [Indexed: 09/25/2023] Open
Abstract
This study investigates the effects of including patients' clinical information on the performance of deep learning (DL) classifiers for disease location in chest X-ray images. Although current classifiers achieve high performance using chest X-ray images alone, consultations with practicing radiologists indicate that clinical data is highly informative and essential for interpreting medical images and making proper diagnoses. In this work, we propose a novel architecture consisting of two fusion methods that enable the model to simultaneously process patients' clinical data (structured data) and chest X-rays (image data). Since these data modalities are in different dimensional spaces, we propose a spatial arrangement strategy, spatialization, to facilitate the multimodal learning process in a Mask R-CNN model. We performed an extensive experimental evaluation using MIMIC-Eye, a dataset comprising different modalities: MIMIC-CXR (chest X-ray images), MIMIC IV-ED (patients' clinical data), and REFLACX (annotations of disease locations in chest X-rays). Results show that incorporating patients' clinical data in a DL model together with the proposed fusion methods improves the disease localization in chest X-rays by 12% in terms of Average Precision compared to a standard Mask R-CNN using chest X-rays alone. Further ablation studies also emphasize the importance of multimodal DL architectures and the incorporation of patients' clinical data in disease localization. In the interest of fostering scientific reproducibility, the architecture proposed within this investigation has been made publicly accessible( https://github.com/ChihchengHsieh/multimodal-abnormalities-detection ).
Collapse
Affiliation(s)
| | | | | | - Chun Ouyang
- Queensland University of Technology, Brisbane, Australia
| | | | - Jacinto C Nascimento
- Institute for Systems and Robotics, Instituto Superior Técnico, University of Lisbon, Lisbon, Portugal
| | - Joaquim Jorge
- Instituto Superior Técnico, University of Lisbon, Portugal, Lisbon, Portugal
| | - Catarina Moreira
- Queensland University of Technology, Brisbane, Australia.
- Instituto Superior Técnico, University of Lisbon, Portugal, Lisbon, Portugal.
- Human Technology Institute, University of Technology Sydney, Ultimo, Australia.
- INESC-ID, Lisbon, Portugal.
| |
Collapse
|
11
|
Wu X, Wang G, Shen N. Research on obstacle avoidance optimization and path planning of autonomous vehicles based on attention mechanism combined with multimodal information decision-making thoughts of robots. Front Neurorobot 2023; 17:1269447. [PMID: 37811356 PMCID: PMC10556461 DOI: 10.3389/fnbot.2023.1269447] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2023] [Accepted: 08/28/2023] [Indexed: 10/10/2023] Open
Abstract
With the development of machine perception and multimodal information decision-making techniques, autonomous driving technology has become a crucial area of advancement in the transportation industry. The optimization of vehicle navigation, path planning, and obstacle avoidance tasks is of paramount importance. In this study, we explore the use of attention mechanisms in a end-to-end architecture for optimizing obstacle avoidance and path planning in autonomous driving vehicles. We position our research within the broader context of robotics, emphasizing the fusion of information and decision-making capabilities. The introduction of attention mechanisms enables vehicles to perceive the environment more accurately by focusing on important information and making informed decisions in complex scenarios. By inputting multimodal information, such as images and LiDAR data, into the attention mechanism module, the system can automatically learn and weigh crucial environmental features, thereby placing greater emphasis on key information during obstacle avoidance decisions. Additionally, we leverage the end-to-end architecture and draw from classical theories and algorithms in the field of robotics to enhance the perception and decision-making abilities of autonomous driving vehicles. Furthermore, we address the optimization of path planning using attention mechanisms. We transform the vehicle's navigation task into a sequential decision-making problem and employ LSTM (Long Short-Term Memory) models to handle dynamic navigation in varying environments. By applying attention mechanisms to weigh key points along the navigation path, the vehicle can flexibly select the optimal route and dynamically adjust it based on real-time conditions. Finally, we conducted extensive experimental evaluations and software experiments on the proposed end-to-end architecture on real road datasets. The method effectively avoids obstacles, adheres to traffic rules, and achieves stable, safe, and efficient autonomous driving in diverse road scenarios. This research provides an effective solution for optimizing obstacle avoidance and path planning in the field of autonomous driving. Moreover, it contributes to the advancement and practical applications of multimodal information fusion in navigation, localization, and human-robot interaction.
Collapse
Affiliation(s)
- Xuejin Wu
- College of Transport and Communications, Shanghai Maritime University, Shanghai, China
| | - Guangming Wang
- School of Management, Wuhan University of Technology, Wuhan, China
- School of Politics and Public Administration, Zhengzhou University, Zhengzhou, China
| | - Nachuan Shen
- Chinese Academy of Fiscal Science, Beijing, China
| |
Collapse
|
12
|
Laghari AA, Sun Y, Alhussein M, Aurangzeb K, Anwar MS, Rashid M. Deep residual-dense network based on bidirectional recurrent neural network for atrial fibrillation detection. Sci Rep 2023; 13:15109. [PMID: 37704659 PMCID: PMC10499947 DOI: 10.1038/s41598-023-40343-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2023] [Accepted: 08/09/2023] [Indexed: 09/15/2023] Open
Abstract
Atrial fibrillation easily leads to stroke, cerebral infarction and other complications, which will seriously harm the life and health of patients. Traditional deep learning methods have weak anti-interference and generalization ability. Therefore, we propose a new-fashioned deep residual-dense network via bidirectional recurrent neural network (RNN) model for atrial fibrillation detection. The combination of one-dimensional dense residual network and bidirectional RNN for atrial fibrillation detection simplifies the tedious feature extraction steps, and constructs the end-to-end neural network to achieve atrial fibrillation detection through data feature learning. Meanwhile, the attention mechanism is utilized to fuse the different features and extract the high-value information. The accuracy of the experimental results is 97.72%, the sensitivity and specificity are 93.09% and 98.71%, respectively compared with other methods.
Collapse
Affiliation(s)
- Asif Ali Laghari
- Software College, Shenyang Normal University, Shenyang, 110034, China
| | - Yanqiu Sun
- Liaoning University of Traditional Chinese Medicine, Shenyang, China.
| | - Musaed Alhussein
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 51178, Riyadh, 11543, Kingdom of Saudi Arabia
| | - Khursheed Aurangzeb
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 51178, Riyadh, 11543, Kingdom of Saudi Arabia
| | | | - Mamoon Rashid
- Department of Computer Engineering, Faculty of Science and Technology, Vishwakarma University, Pune, 411048, India
| |
Collapse
|
13
|
Innani S, Dutande P, Baid U, Pokuri V, Bakas S, Talbar S, Baheti B, Guntuku SC. Generative adversarial networks based skin lesion segmentation. Sci Rep 2023; 13:13467. [PMID: 37596306 PMCID: PMC10439152 DOI: 10.1038/s41598-023-39648-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2023] [Accepted: 07/28/2023] [Indexed: 08/20/2023] Open
Abstract
Skin cancer is a serious condition that requires accurate diagnosis and treatment. One way to assist clinicians in this task is using computer-aided diagnosis tools that automatically segment skin lesions from dermoscopic images. We propose a novel adversarial learning-based framework called Efficient-GAN (EGAN) that uses an unsupervised generative network to generate accurate lesion masks. It consists of a generator module with a top-down squeeze excitation-based compound scaled path, an asymmetric lateral connection-based bottom-up path, and a discriminator module that distinguishes between original and synthetic masks. A morphology-based smoothing loss is also implemented to encourage the network to create smooth semantic boundaries of lesions. The framework is evaluated on the International Skin Imaging Collaboration Lesion Dataset. It outperforms the current state-of-the-art skin lesion segmentation approaches with a Dice coefficient, Jaccard similarity, and accuracy of 90.1%, 83.6%, and 94.5%, respectively. We also design a lightweight segmentation framework called Mobile-GAN (MGAN) that achieves comparable performance as EGAN but with an order of magnitude lower number of training parameters, thus resulting in faster inference times for low compute resource settings.
Collapse
Affiliation(s)
- Shubham Innani
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, Maharashtra, India.
| | - Prasad Dutande
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, Maharashtra, India
| | - Ujjwal Baid
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, Maharashtra, India
- Center for Biomedical Image Computing and Analytics, University of Pennsylvania, Philadelphia, USA
| | - Venu Pokuri
- Department of Computer and Information Science, University of Pennsylvania, Philadelphia, USA
| | - Spyridon Bakas
- Center for Biomedical Image Computing and Analytics, University of Pennsylvania, Philadelphia, USA
| | - Sanjay Talbar
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, Maharashtra, India
| | - Bhakti Baheti
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, Maharashtra, India
- Center for Biomedical Image Computing and Analytics, University of Pennsylvania, Philadelphia, USA
| | - Sharath Chandra Guntuku
- Department of Computer and Information Science, University of Pennsylvania, Philadelphia, USA
| |
Collapse
|
14
|
Johnson DR, Vaidhyanathan RU. Detection and localization of multi-scale and oriented objects using an enhanced feature refinement algorithm. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:15219-15243. [PMID: 37679178 DOI: 10.3934/mbe.2023681] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/09/2023]
Abstract
Object detection is a fundamental aspect of computer vision, with numerous generic object detectors proposed by various researchers. The proposed work presents a novel single-stage rotation detector that can detect oriented and multi-scale objects accurately from diverse scenarios. This detector addresses the challenges faced by current rotation detectors, such as the detection of arbitrary orientations, objects that are densely arranged, and the issue of loss discontinuity. First, the detector also adopts a progressive regression form (coarse-to-fine-grained approach) that uses both horizontal anchors (speed and higher recall) and rotating anchors (oriented objects) in cluttered backgrounds. Second, the proposed detector includes a feature refinement module that helps minimize the problems related to feature angulation and reduces the number of bounding boxes generated. Finally, to address the issue of loss discontinuity, the proposed detector utilizes a newly formulated adjustable loss function that can be extended to both single-stage and two-stage detectors. The proposed detector shows outstanding performance on benchmark datasets and significantly outperforms other state-of-the-art methods in terms of speed and accuracy.
Collapse
Affiliation(s)
- Deepika Roselind Johnson
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamilnadu, India
| | | |
Collapse
|
15
|
Zhang Y, Yuan Q, Muzzammil HM, Gao G, Xu Y. Image-guided prostate biopsy robots: A review. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:15135-15166. [PMID: 37679175 DOI: 10.3934/mbe.2023678] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/09/2023]
Abstract
At present, the incidence of prostate cancer (PCa) in men is increasing year by year. So, the early diagnosis of PCa is of great significance. Transrectal ultrasonography (TRUS)-guided biopsy is a common method for diagnosing PCa. The biopsy process is performed manually by urologists but the diagnostic rate is only 20%-30% and its reliability and accuracy can no longer meet clinical needs. The image-guided prostate biopsy robot has the advantages of a high degree of automation, does not rely on the skills and experience of operators, reduces the work intensity and operation time of urologists and so on. Capable of delivering biopsy needles to pre-defined biopsy locations with minimal needle placement errors, it makes up for the shortcomings of traditional free-hand biopsy and improves the reliability and accuracy of biopsy. The integration of medical imaging technology and the robotic system is an important means for accurate tumor location, biopsy puncture path planning and visualization. This paper mainly reviews image-guided prostate biopsy robots. According to the existing literature, guidance modalities are divided into magnetic resonance imaging (MRI), ultrasound (US) and fusion image. First, the robot structure research by different guided methods is the main line and the actuators and material research of these guided modalities is the auxiliary line to introduce and compare. Second, the robot image-guided localization technology is discussed. Finally, the image-guided prostate biopsy robot is summarized and suggestions for future development are provided.
Collapse
Affiliation(s)
- Yongde Zhang
- Key Laboratory of Advanced Manufacturing and Intelligent Technology, Ministry of Education, Harbin University of Science and Technology, Harbin 150080, China
- Foshan Baikang Robot Technology Co., Ltd, Nanhai District, Foshan City, Guangdong Province 528225, China
| | - Qihang Yuan
- Key Laboratory of Advanced Manufacturing and Intelligent Technology, Ministry of Education, Harbin University of Science and Technology, Harbin 150080, China
| | - Hafiz Muhammad Muzzammil
- Key Laboratory of Advanced Manufacturing and Intelligent Technology, Ministry of Education, Harbin University of Science and Technology, Harbin 150080, China
| | - Guoqiang Gao
- Key Laboratory of Advanced Manufacturing and Intelligent Technology, Ministry of Education, Harbin University of Science and Technology, Harbin 150080, China
| | - Yong Xu
- Department of Urology, the Third Medical Centre, Chinese PLA (People's Liberation Army) General Hospital, Beijing 100039, China
| |
Collapse
|
16
|
Yoshida Y, Hayashi Y, Shimada T, Hattori N, Tomita K, Miura RE, Yamao Y, Tateishi S, Iwadate Y, Nakada TA. Prehospital stroke-scale machine-learning model predicts the need for surgical intervention. Sci Rep 2023; 13:9135. [PMID: 37277424 DOI: 10.1038/s41598-023-36004-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 05/27/2023] [Indexed: 06/07/2023] Open
Abstract
While the development of prehospital diagnosis scales has been reported in various regions, we have also developed a scale to predict stroke type using machine learning. In the present study, we aimed to assess for the first time a scale that predicts the need for surgical intervention across stroke types, including subarachnoid haemorrhage and intracerebral haemorrhage. A multicentre retrospective study was conducted within a secondary medical care area. Twenty-three items, including vitals and neurological symptoms, were analysed in adult patients suspected of having a stroke by paramedics. The primary outcome was a binary classification model for predicting surgical intervention based on eXtreme Gradient Boosting (XGBoost). Of the 1143 patients enrolled, 765 (70%) were used as the training cohort, and 378 (30%) were used as the test cohort. The XGBoost model predicted stroke requiring surgical intervention with high accuracy in the test cohort, with an area under the receiver operating characteristic curve of 0.802 (sensitivity 0.748, specificity 0.853). We found that simple survey items, such as the level of consciousness, vital signs, sudden headache, and speech abnormalities were the most significant variables for accurate prediction. This algorithm can be useful for prehospital stroke management, which is crucial for better patient outcomes.
Collapse
Affiliation(s)
- Yoichi Yoshida
- Department of Neurosurgery, Chiba Municipal Kaihin Hospital, Chiba, Japan
- Department of Neurological Surgery, Graduate School of Medicine, Chiba University, Chiba, Japan
| | - Yosuke Hayashi
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
| | - Tadanaga Shimada
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
| | - Noriyuki Hattori
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
| | - Keisuke Tomita
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
| | - Rie E Miura
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
- SMART119 Inc., 7th Floor, Chiba Chuo Twin Building No. 2, 2-5-1 Chuo, Chiba, Japan
| | - Yasuo Yamao
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
- SMART119 Inc., 7th Floor, Chiba Chuo Twin Building No. 2, 2-5-1 Chuo, Chiba, Japan
| | - Shino Tateishi
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan
- SMART119 Inc., 7th Floor, Chiba Chuo Twin Building No. 2, 2-5-1 Chuo, Chiba, Japan
| | - Yasuo Iwadate
- Department of Neurological Surgery, Graduate School of Medicine, Chiba University, Chiba, Japan
| | - Taka-Aki Nakada
- Department of Emergency and Critical Care Medicine, Chiba University Graduate School of Medicine, 1-8-1 Inohana, Chuo, Chiba, 260-8677, Japan.
- SMART119 Inc., 7th Floor, Chiba Chuo Twin Building No. 2, 2-5-1 Chuo, Chiba, Japan.
| |
Collapse
|
17
|
Liu P, Zheng G. CVCL: Context-aware Voxel-wise Contrastive Learning for label-efficient multi-organ segmentation. Comput Biol Med 2023; 160:106995. [PMID: 37187134 DOI: 10.1016/j.compbiomed.2023.106995] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2022] [Revised: 04/02/2023] [Accepted: 05/01/2023] [Indexed: 05/17/2023]
Abstract
Despite the significant performance improvement on multi-organ segmentation with supervised deep learning-based methods, the label-hungry nature hinders their applications in practical disease diagnosis and treatment planning. Due to the challenges in obtaining expert-level accurate, densely annotated multi-organ datasets, label-efficient segmentation, such as partially supervised segmentation trained on partially labeled datasets or semi-supervised medical image segmentation, has attracted increasing attention recently. However, most of these methods suffer from the limitation that they neglect or underestimate the challenging unlabeled regions during model training. To this end, we propose a novel Context-aware Voxel-wise Contrastive Learning method, referred as CVCL, to take full advantage of both labeled and unlabeled information in label-scarce datasets for a performance improvement on multi-organ segmentation. Experimental results demonstrate that our proposed method achieves superior performance than other state-of-the-art methods.
Collapse
Affiliation(s)
- Peng Liu
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, No. 800, Dongchuan Road, Shanghai, 200240, China
| | - Guoyan Zheng
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, No. 800, Dongchuan Road, Shanghai, 200240, China.
| |
Collapse
|
18
|
Kasbekar RS, Ji S, Clancy EA, Goel A. Optimizing the input feature sets and machine learning algorithms for reliable and accurate estimation of continuous, cuffless blood pressure. Sci Rep 2023; 13:7750. [PMID: 37173370 PMCID: PMC10181996 DOI: 10.1038/s41598-023-34677-9] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2023] [Accepted: 05/05/2023] [Indexed: 05/15/2023] Open
Abstract
The advent of mobile devices, wearables and digital healthcare has unleashed a demand for accurate, reliable, and non-interventional ways to measure continuous blood pressure (BP). Many consumer products claim to measure BP with a cuffless device, but their lack of accuracy and reliability limit clinical adoption. Here, we demonstrate how multimodal feature datasets, comprising: (i) pulse arrival time (PAT); (ii) pulse wave morphology (PWM), and (iii) demographic data, can be combined with optimized Machine Learning (ML) algorithms to estimate Systolic BP (SBP), Diastolic BP (DBP) and Mean Arterial Pressure (MAP) within a 5 mmHg bias of the gold standard Intra-Arterial BP, well within the acceptable limits of the IEC/ANSI 80601-2-30 (2018) standard. Furthermore, DBP's calculated using 126 datasets collected from 31 hemodynamically compromised patients had a standard deviation within 8 mmHg, while SBP's and MAP's exceeded these limits. Using ANOVA and Levene's test for error means and standard deviations, we found significant differences in the various ML algorithms but found no significant differences amongst the multimodal feature datasets. Optimized ML algorithms and key multimodal features obtained from larger real-world data (RWD) sets could enable more reliable and accurate estimation of continuous BP in cuffless devices, accelerating wider clinical adoption.
Collapse
Affiliation(s)
- Rajesh S Kasbekar
- Department of Biomedical Engineering, Worcester Polytechnic Institute (WPI), Worcester, MA, USA.
| | - Songbai Ji
- Department of Biomedical Engineering, Worcester Polytechnic Institute (WPI), Worcester, MA, USA
| | - Edward A Clancy
- Department of Biomedical Engineering, Worcester Polytechnic Institute (WPI), Worcester, MA, USA
- Department of Electrical and Computer Engineering, Worcester Polytechnic Institute (WPI), Worcester, MA, USA
| | - Anita Goel
- Nanobiosym Research Institute, Nanobiosym, Inc. and Department of Physics, Harvard University, Cambridge, MA, USA
| |
Collapse
|
19
|
Hwang JH, Lim M, Han G, Park H, Kim YB, Park J, Jun SY, Lee J, Cho JW. Preparing pathological data to develop an artificial intelligence model in the nonclinical study. Sci Rep 2023; 13:3896. [PMID: 36890209 PMCID: PMC9994413 DOI: 10.1038/s41598-023-30944-x] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Accepted: 03/03/2023] [Indexed: 03/10/2023] Open
Abstract
Artificial intelligence (AI)-based analysis has recently been adopted in the examination of histological slides via the digitization of glass slides using a digital scanner. In this study, we examined the effect of varying the staining color tone and magnification level of a dataset on the result of AI model prediction in hematoxylin and eosin stained whole slide images (WSIs). The WSIs of liver tissues with fibrosis were used as an example, and three different datasets (N20, B20, and B10) were prepared with different color tones and magnifications. Using these datasets, we built five models trained Mask R-CNN algorithm by a single or mixed dataset of N20, B20, and B10. We evaluated their model performance using the test dataset of three datasets. It was found that the models that were trained with mixed datasets (models B20/N20 and B10/B20), which consist of different color tones or magnifications, performed better than the single dataset trained models. Consequently, superior performance of the mixed models was obtained from the actual prediction results of the test images. We suggest that training the algorithm with various staining color tones and multi-scaled image datasets would be more optimized for consistent remarkable performance in predicting pathological lesions of interest.
Collapse
Affiliation(s)
- Ji-Hee Hwang
- Toxicologic Pathology Research Group, Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea
| | - Minyoung Lim
- Toxicologic Pathology Research Group, Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea
| | - Gyeongjin Han
- Toxicologic Pathology Research Group, Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea
| | - Heejin Park
- Toxicologic Pathology Research Group, Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea
| | - Yong-Bum Kim
- Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea
| | - Jinseok Park
- Research and Development Team, LAC Inc., Seoul, 07807, Korea
| | - Sang-Yeop Jun
- Research and Development Team, LAC Inc., Seoul, 07807, Korea
| | - Jaeku Lee
- Research and Development Team, LAC Inc., Seoul, 07807, Korea
| | - Jae-Woo Cho
- Toxicologic Pathology Research Group, Department of Advanced Toxicology Research, Korea Institute of Toxicology, Daejeon, 34114, Korea.
| |
Collapse
|
20
|
Liu M, Wang Z, Li H, Wu P, Alsaadi FE, Zeng N. AA-WGAN: Attention augmented Wasserstein generative adversarial network with application to fundus retinal vessel segmentation. Comput Biol Med 2023; 158:106874. [PMID: 37019013 DOI: 10.1016/j.compbiomed.2023.106874] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2023] [Revised: 03/15/2023] [Accepted: 03/30/2023] [Indexed: 04/03/2023]
Abstract
In this paper, a novel attention augmented Wasserstein generative adversarial network (AA-WGAN) is proposed for fundus retinal vessel segmentation, where a U-shaped network with attention augmented convolution and squeeze-excitation module is designed to serve as the generator. In particular, the complex vascular structures make some tiny vessels hard to segment, while the proposed AA-WGAN can effectively handle such imperfect data property, which is competent in capturing the dependency among pixels in the whole image to highlight the regions of interests via the applied attention augmented convolution. By applying the squeeze-excitation module, the generator is able to pay attention to the important channels of the feature maps, and the useless information can be suppressed as well. In addition, gradient penalty method is adopted in the WGAN backbone to alleviate the phenomenon of generating large amounts of repeated images due to excessive concentration on accuracy. The proposed model is comprehensively evaluated on three datasets DRIVE, STARE, and CHASE_DB1, and the results show that the proposed AA-WGAN is a competitive vessel segmentation model as compared with several other advanced models, which obtains the accuracy of 96.51%, 97.19% and 96.94% on each dataset, respectively. The effectiveness of the applied important components is validated by ablation study, which also endows the proposed AA-WGAN with considerable generalization ability.
Collapse
|