1
|
Kusuma PV, Reddy SCM. Brain tumor segmentation and classification using MRI: Modified segnet model and hybrid deep learning architecture with improved texture features. Comput Biol Chem 2025; 117:108381. [PMID: 40020564 DOI: 10.1016/j.compbiolchem.2025.108381] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2024] [Revised: 01/29/2025] [Accepted: 02/11/2025] [Indexed: 03/03/2025]
Abstract
Brain tumors are quickly overtaking all other causes of death worldwide. The failure to perform a timely diagnosis is the main cause of increasing the death rate. Traditional methods of brain tumor diagnosis heavily rely on the expertise of radiologists, making timely and accurate diagnosis challenging. Magnetic Resonance Imaging (MRI) has emerged as the primary modality for brain tumor detection, but manual interpretation of MRI scans is time-consuming and error-prone. To address these challenges, an automated approach for brain tumor segmentation and classification (BTS&C) using MRI scans is proposed in this work. This work suggests a brain tumor classification scheme using MRI. Initially, the input images T1, TIC, t2 and t2 flair are fused via an improved fusion method. Then, Median Filtering (MF) is applied to preprocess the fused image. Also, the Modified Segnet model is proposed with a new pooling operation to do the segmentation process. Features like Improved local Gabor Binay pattern Histogram Sequence (ILGBPHS), Weber Local descriptor (WLD), and Tetrolet waveform are extracted from the segmented image. Finally, classification is done with HDLA that combines Bi-LSTM and Modified Linknet models. When TD= 90 %, the proposed method achieves a higher accuracy of 98 % which is compared to other methods like Bi-LSTM, Link Net, LeNet, Squeeze Net, Efficient Net, HHOCNN and CNN-SVM.
Collapse
Affiliation(s)
- Palleti Venkata Kusuma
- Department of Electronics and Communication Engineering Jawaharlal Nehru Technological University Anantapur, Ananthapuramu, Andhra Pradesh 515002, India.
| | - S Chandra Mohan Reddy
- Department of Electronics and Communication Engineering Jawaharlal Nehru Technological University Anantapur, Ananthapuramu, Andhra Pradesh 515002, India.
| |
Collapse
|
2
|
Xu H, Lv R. Rapid diagnosis of lung cancer by multi-modal spectral data combined with deep learning. SPECTROCHIMICA ACTA. PART A, MOLECULAR AND BIOMOLECULAR SPECTROSCOPY 2025; 335:125997. [PMID: 40073660 DOI: 10.1016/j.saa.2025.125997] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/09/2024] [Revised: 02/21/2025] [Accepted: 03/04/2025] [Indexed: 03/14/2025]
Abstract
Lung cancer is a malignant tumor that poses a serious threat to human health. Existing lung cancer diagnostic techniques face the challenges of high cost and slow diagnosis. Early and rapid diagnosis and treatment are essential to improve the outcome of lung cancer. In this study, a deep learning-based multi-modal spectral information fusion (MSIF) network is proposed for lung adenocarcinoma cell detection. First, multi-modal data of Fourier transform infrared spectra, UV-vis absorbance spectra, and fluorescence spectra of normal and patient cells were collected. Subsequently, the spectral text data were efficiently processed by one-dimensional convolutional neural network. The global and local features of the spectral images are deeply mined by the hybrid model of ResNet and Transformer. An adaptive depth-wise convolution (ADConv) is introduced to be applied to feature extraction, overcoming the shortcomings of conventional convolution. In order to achieve feature learning between multi-modalities, a cross-modal interaction fusion (CMIF) module is designed. This module fuses the extracted spectral image and text features in a multi-faceted interaction, enabling full utilization of multi-modal features through feature sharing. The method demonstrated excellent performance on the test sets of Fourier transform infrared spectra, UV-vis absorbance spectra and fluorescence spectra, achieving 95.83 %, 97.92 % and 100 % accuracy, respectively. In addition, experiments validate the superiority of multi-modal spectral data and the robustness of the model generalization capability. This study not only provides strong technical support for the early diagnosis of lung cancer, but also opens a new chapter for the application of multi-modal data fusion in spectroscopy.
Collapse
Affiliation(s)
- Han Xu
- State Key Laboratory of Electromechanical Integrated Manufacturing of High-performance Electronic Equipment, School of Electro-Mechanical Engineering, Xidian University, Xi'an, Shaanxi 710071, China
| | - Ruichan Lv
- State Key Laboratory of Electromechanical Integrated Manufacturing of High-performance Electronic Equipment, School of Electro-Mechanical Engineering, Xidian University, Xi'an, Shaanxi 710071, China.
| |
Collapse
|
3
|
Reddy C KK, Daduvy A, Kaza VS, Shuaib M, Mohzary M, Alam S, Sheneamer A. A multi-scale convolutional LSTM-dense network for robust cardiac arrhythmia classification from ECG signals. Comput Biol Med 2025; 191:110121. [PMID: 40233677 DOI: 10.1016/j.compbiomed.2025.110121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2024] [Revised: 03/28/2025] [Accepted: 03/31/2025] [Indexed: 04/17/2025]
Abstract
Cardiac arrhythmias are irregular heart rhythms that, if undetected, can lead to severe cardiovascular conditions. Detecting these anomalies early through electrocardiogram (ECG) signal analysis is critical for preventive healthcare and effective treatment. However, the automatic classification of arrhythmias poses significant challenges, including class imbalance and noise interference in ECG signals. This paper introduces the Multi-Scale Convolutional LSTM Dense Network (MS-CLDNet) model, an advanced deep-learning model specifically designed to address these issues and improve arrhythmia classification accuracy and other relevant metrics. This paper aims to develop an efficient deep-learning model, MS-CLDNet, for accurately classifying cardiac arrhythmias from electrocardiogram (ECG) signals. Addressing challenges like class imbalance and noise interference, the model integrates bidirectional long short-term memory (LSTM) networks for temporal pattern recognition, Dense Blocks for feature refinement, and Multi-Scale Convolutional Neural Networks (CNNs) for robust feature extraction. To achieve accurate classification of different types of arrhythmias, the Classification Head refines these extracted features even further. Utilizing the MIT-BIH arrhythmia dataset, key pre-processing techniques such as wavelet-based denoising were employed to enhance signal clarity. Results indicate that the MS-CLDNet model achieves a classification accuracy of 98.22 %, outperforming baseline models with low average loss values (0.084). This research highlights how crucial it is to combine sophisticated neural network architectures with efficient pre-processing techniques to improve the precision and accuracy of automated cardiovascular diagnostic systems, which could have important healthcare applications for early and accurate arrhythmia detection.
Collapse
Affiliation(s)
- Kishor Kumar Reddy C
- Department of Computer Science and Engineering, Stanley College of Engineering and Technology for Women, Hyderabad, Telangana, India
| | - Advaitha Daduvy
- Department of Computer Science and Engineering, Stanley College of Engineering and Technology for Women, Hyderabad, Telangana, India
| | - Vijaya Sindhoori Kaza
- Department of Computer Science and Engineering, Stanley College of Engineering and Technology for Women, Hyderabad, Telangana, India
| | - Mohammed Shuaib
- Department of Computer Science, College of Engineering and Computer Science, Jazan University, Jazan, 45142, Saudi Arabia
| | - Muhammad Mohzary
- Department of Computer Science, College of Engineering and Computer Science, Jazan University, Jazan, 45142, Saudi Arabia; Engineering and Technology Research Center, Jazan University, P.O. Box 114, Jazan, 82817, Saudi Arabia.
| | - Shadab Alam
- Department of Computer Science, College of Engineering and Computer Science, Jazan University, Jazan, 45142, Saudi Arabia
| | - Abdullah Sheneamer
- Department of Computer Science, College of Engineering and Computer Science, Jazan University, Jazan, 45142, Saudi Arabia
| |
Collapse
|
4
|
Missaoui R, Hechkel W, Saadaoui W, Helali A, Leo M. Advanced Deep Learning and Machine Learning Techniques for MRI Brain Tumor Analysis: A Review. SENSORS (BASEL, SWITZERLAND) 2025; 25:2746. [PMID: 40363185 PMCID: PMC12074157 DOI: 10.3390/s25092746] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/21/2025] [Revised: 04/09/2025] [Accepted: 04/23/2025] [Indexed: 05/15/2025]
Abstract
A brain tumor is the result of abnormal growth of cells in the central nervous system (CNS), widely considered as a complex and diverse clinical entity that is difficult to diagnose and cure. In this study, we focus on current advances in medical imaging, particularly magnetic resonance imaging (MRI), and how machine learning (ML) and deep learning (DL) algorithms might be combined with clinical assessments to improve brain tumor diagnosis. Due to its superior contrast resolution and safety compared to other imaging methods, MRI is highlighted as the preferred imaging modality for brain tumors. The challenges related to brain tumor analysis in different processes including detection, segmentation, classification, and survival prediction are addressed along with how ML/DL approaches significantly improve these steps. We systematically analyzed 107 studies (2018-2024) employing ML, DL, and hybrid models across publicly available datasets such as BraTS, TCIA, and Figshare. In the light of recent developments in brain tumor analysis, many algorithms have been proposed to accurately obtain ontological characteristics of tumors, enhancing diagnostic precision and personalized therapeutic strategies.
Collapse
Affiliation(s)
- Rim Missaoui
- Laboratory of Micro-Optoelectronics and Nanostructures (LMON), University of Monastir, Avenue of the Environment, Monastir 5019, Tunisia; (R.M.); (W.H.); (A.H.)
- National High School of Engineering of Tunis (ENSIT), 5 Rue Taha Hussein–Montfleury, Tunis 1008, Tunisia
| | - Wided Hechkel
- Laboratory of Micro-Optoelectronics and Nanostructures (LMON), University of Monastir, Avenue of the Environment, Monastir 5019, Tunisia; (R.M.); (W.H.); (A.H.)
| | - Wajdi Saadaoui
- LRMAN Laboratory, Higher Institute of Applied Sciences and Technology of Kasserine (ISSAT), Kasserine 1200, Tunisia;
| | - Abdelhamid Helali
- Laboratory of Micro-Optoelectronics and Nanostructures (LMON), University of Monastir, Avenue of the Environment, Monastir 5019, Tunisia; (R.M.); (W.H.); (A.H.)
| | - Marco Leo
- Institute of Applied Sciences and Intelligent Systems, National Research Council of Italy, 73100 Lecce, Italy
| |
Collapse
|
5
|
D N S, Pai RM, Bhat SN, Pai MMM. Vision transformer and deep learning based weighted ensemble model for automated spine fracture type identification with GAN generated CT images. Sci Rep 2025; 15:14408. [PMID: 40274849 PMCID: PMC12022092 DOI: 10.1038/s41598-025-98518-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2024] [Accepted: 04/11/2025] [Indexed: 04/26/2025] Open
Abstract
The most common causes of spine fractures, or vertebral column fractures (VCF), are traumas like falls, injuries from sports, or accidents. CT scans are affordable and effective at detecting VCF types in an accurate manner. VCF type identification in cervical, thoracic, and lumbar (C3-L5) regions is limited and sensitive to inter-observer variability. To solve this problem, this work introduces an autonomous approach for identifying VCF type by developing a novel ensemble model of Vision Transformers (ViT) and best-performing deep learning (DL) models. It assists orthopaedicians in easy and early identification of VCF types. The performance of numerous fine-tuned DL architectures, including VGG16, ResNet50, and DenseNet121, was investigated, and an ensemble classification model was developed to identify the best-performing combination of DL models. A ViT model is also trained to identify VCF. Later, the best-performing DL models and ViT were fused by weighted average technique for type identification. To overcome data limitations, an extended Deep Convolutional Generative Adversarial Network (DCGAN) and Progressive Growing Generative Adversarial Network (PGGAN) were developed. The VGG16-ResNet50-ViT ensemble model outperformed all ensemble models and got an accuracy of 89.98%. Extended DCGAN and PGGAN augmentation increased the accuracy of type identification to 90.28% and 93.68%, respectively. This demonstrates efficacy of PGGANs in augmenting VCF images. The study emphasizes the distinctive contributions of the ResNet50, VGG16, and ViT models in feature extraction, generalization, and global shape-based pattern capturing in VCF type identification. CT scans collected from a tertiary care hospital are used to validate these models.
Collapse
Affiliation(s)
- Sindhura D N
- Department of Data Science and Computer Applications, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, India
| | - Radhika M Pai
- Department of Data Science and Computer Applications, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, India.
| | - Shyamasunder N Bhat
- Department of Orthopaedics, Kasturba Medical College, Manipal Academy of Higher Education, Manipal, India
| | - Manohara M M Pai
- Department of Information and Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, India
| |
Collapse
|
6
|
Mohanty MR, Mallick PK, Mishra D. Bald eagle-optimized transformer networks with temporal-spatial mid-level features for pancreatic tumor classification. Biomed Phys Eng Express 2025; 11:035019. [PMID: 40203846 DOI: 10.1088/2057-1976/adcac9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2025] [Accepted: 04/09/2025] [Indexed: 04/11/2025]
Abstract
The classification and diagnosis of pancreatic tumors present significant challenges due to their inherent complexity and variability. Traditional methods often struggle to capture the dynamic nature of these tumors, highlighting the need for advanced techniques that improve precision and robustness. This study introduces a novel approach that combines temporal-spatial mid-level features (CTSF) with bald eagle search (BES) optimized transformer networks to enhance pancreatic tumor classification. By leveraging temporal-spatial features that encompass both spatial structure and temporal evolution, we employ the BES algorithm to optimize the vision transformer (ViT) and swin transformer (ST) models, significantly enhancing their capacity to process complex datasets. The study underscores the critical role of temporal features in pancreatic tumor classification, enabling the capture of changes over time to improve our understanding of tumor progression and treatment responses. Among the models evaluated-GRU, LSTM, and ViT-the ViTachieved superior performance, with accuracy rates of 94.44%, 89.44%, and 87.22% on the TCIA-Pancreas-CT, Decathlon Pancreas, and NIH-Pancreas-CT datasets, respectively. Spatial features extracted from ResNet50, VGG16, and ST were also essential, with the ST model attaining the highest accuracy of 95.00%, 95.56%, and 93.33% on the same datasets. The integration of temporal and spatial features within the CTSF model resulted in accuracy rates of 96.02%, 97.21%, and 95.06% for the TCIA-Pancreas-CT, Decathlon Pancreas, and NIH-Pancreas-CT datasets, respectively. Furthermore, optimization techniques, particularly hyperparameter tuning, further enhanced performance, with the BES-optimized model achieving the highest accuracy of 98.02%, 98.92%, and 98.89%. The superiority of the CTSF-BES approach was confirmed through the Friedman test and Bonferroni-Dunn test, while execution time analysis demonstrated a favourable balance between performance and efficiency.
Collapse
Affiliation(s)
- Manas Ranjan Mohanty
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar, Odisha, India
| | - Pradeep Kumar Mallick
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar, Odisha, India
| | - Debahuti Mishra
- Department of Computer Science and Engineering, Siksha 'O' Anusandhan (Deemed to be) University, Bhubaneswar, Odisha, India
| |
Collapse
|
7
|
Abirami G, Nagadevi S, Dorathi Jayaseeli JD, Rao TP, Patibandla RSML, Aluvalu R, Srihari K. An integration of ensemble deep learning with hybrid optimization approaches for effective underwater object detection and classification model. Sci Rep 2025; 15:10902. [PMID: 40158003 PMCID: PMC11954852 DOI: 10.1038/s41598-025-95596-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2024] [Accepted: 03/21/2025] [Indexed: 04/01/2025] Open
Abstract
Underwater object detection (UOD) is essential in maritime environmental study and underwater species protection. The development of associated technology holds real-world importance. While current object recognition methods have attained an outstanding performance on terrestrial, they are less suitable in underwater conditions because of dual restrictions: the underwater objects are generally smaller, closely spread, and disposed to obstruction features, and underwater embedding tools have temporary storing and computation abilities. Image-based UOD has progressed fast recently, in addition to deep learning (DL) applications and development in computer vision (CV). Investigators utilize DL models to identify possible objects inside an image. Convolutional neural network (CNN) is the major technique of DL, which enhances the learning qualities. In this manuscript, an Underwater Object Detection and Classification Utilizing the Ensemble Deep Learning Approach and Hybrid Optimization Algorithms (UODC-EDLHOA) technique is developed. The UODC-EDLHOA technique mainly detects and classifies underwater objects using advanced DL and hyperparameter models. Initially, the UODC-EDLHOA model involved several levels of pre-processing and noise removal to improve the clearness of the underwater images. The backbone of EfficientNetB7, which has an attention mechanism, is employed for feature extraction. Furthermore, the YOLOv9-based object detection is utilized. For underwater object detection, an ensemble of three techniques, namely deep neural network (DNN), deep belief network (DBN), and long short-term memory (LSTM), is implemented. Finally, the hyperparameter selection uses the hybrid Siberian tiger and sand cat swarm optimization (STSC) methods. Extensive experimentation is conducted on the UOD dataset to illustrate the robust classification performance of the UODC-EDLHOA model. The performance validation of the UODC-EDLHOA model portrayed a superior accuracy value of 92.78% over existing techniques.
Collapse
Affiliation(s)
- G Abirami
- Department of Computing Technologies, School of Computing, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - S Nagadevi
- Department of Computing Technologies, School of Computing, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - J D Dorathi Jayaseeli
- Department of Computing Technologies, School of Computing, SRM Institute of Science and Technology, Kattankulathur, 603203, Chennai, India
| | - T Prabhakara Rao
- Department of Computer Science and Engineering, Aditya University, Surampalem, Andhra Pradesh, India
| | - R S M Lakshmi Patibandla
- Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Guntur, Andhra Pradesh, India
| | - Rajanikanth Aluvalu
- Symbiosis Institute of Technology, Hyderabad Campus, Symbiosis International (Deemed University), Pune, India.
| | - K Srihari
- Department of Computer Science and Engineering, SNS College of Technology, Coimbatore, India
| |
Collapse
|
8
|
Li J, Niu Y, Du J, Wu J, Guo W, Wang Y, Wang J, Mu J. HTRecNet: a deep learning study for efficient and accurate diagnosis of hepatocellular carcinoma and cholangiocarcinoma. Front Cell Dev Biol 2025; 13:1549811. [PMID: 40196844 PMCID: PMC11973358 DOI: 10.3389/fcell.2025.1549811] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2024] [Accepted: 03/03/2025] [Indexed: 04/09/2025] Open
Abstract
Background Hepatocellular carcinoma (HCC) and cholangiocarcinoma (CCA) represent the primary liver cancer types. Traditional diagnostic techniques, reliant on radiologist interpretation, are both time-intensive and often inadequate for detecting the less prevalent CCA. There is an emergent need to explore automated diagnostic methods using deep learning to address these challenges. Methods This study introduces HTRecNet, a novel deep learning framework for enhanced diagnostic precision and efficiency. The model incorporates sophisticated data augmentation strategies to optimize feature extraction, ensuring robust performance even with constrained sample sizes. A comprehensive dataset of 5,432 histopathological images was divided into 5,096 for training and validation, and 336 for external testing. Evaluation was conducted using five-fold cross-validation and external validation, applying metrics such as accuracy, area under the receiver operating characteristic curve (AUC), and Matthews correlation coefficient (MCC) against established clinical benchmarks. Results The training and validation cohorts comprised 1,536 images of normal liver tissue, 3,380 of HCC, and 180 of CCA. HTRecNet showed exceptional efficacy, consistently achieving AUC values over 0.99 across all categories. In external testing, the model reached an accuracy of 0.97 and an MCC of 0.95, affirming its reliability in distinguishing between normal, HCC, and CCA tissues. Conclusion HTRecNet markedly enhances the capability for early and accurate differentiation of HCC and CCA from normal liver tissues. Its high diagnostic accuracy and efficiency position it as an invaluable tool in clinical settings, potentially transforming liver cancer diagnostic protocols. This system offers substantial support for refining diagnostic workflows in healthcare environments focused on liver malignancies.
Collapse
Affiliation(s)
- Jingze Li
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
| | - Yupeng Niu
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
- Artificial Intelligence Laboratory, Sichuan Agricultural University, Ya’ an, China
| | - Junwu Du
- Department of Hepatobiliary Pancreaticosplenic Surgery, Ya ‘an People’s Hospital, Ya’ an, China
| | - Jiani Wu
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
| | - Weichen Guo
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
- Artificial Intelligence Laboratory, Sichuan Agricultural University, Ya’ an, China
| | - Yujie Wang
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
| | - Jian Wang
- Department of Neurology, Ya’an People’s Hospital, Ya’ an, China
- Department of Neurology, The Affiliated Hospital, Southwest Medical University, Luzhou, China
| | - Jiong Mu
- College of Information Engineering, Sichuan Agricultural University, Ya’ an, China
- Artificial Intelligence Laboratory, Sichuan Agricultural University, Ya’ an, China
| |
Collapse
|
9
|
Choi JC, Kim YJ, Kim KG, Kim EY. An Analysis of the Efficacy of Deep Learning-Based Pectoralis Muscle Segmentation in Chest CT for Sarcopenia Diagnosis. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01443-4. [PMID: 40011347 DOI: 10.1007/s10278-025-01443-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/16/2024] [Revised: 02/04/2025] [Accepted: 02/05/2025] [Indexed: 02/28/2025]
Abstract
Sarcopenia is the loss of skeletal muscle function and mass and is a poor prognostic factor. This condition is typically diagnosed by measuring skeletal muscle mass at the L3 level. Chest computed tomography (CT) scans do not include the L3 level. We aimed to determine if these scans can be used to diagnose sarcopenia and thus guide patient management and treatment decisions. This study compared the ResNet-UNet, Recurrent Residual UNet, and UNet3 + models for segmenting and measuring the pectoralis muscle area in chest CT images. A total of 4932 chest CT images were collected from 1644 patients, and additional abdominal CT data were collected from 294 patients. The performance of the models was evaluated using the dice similarity coefficient (DSC), accuracy, sensitivity, and specificity. Furthermore, the correlation between the segmented pectoralis and L3 muscle areas was compared using linear regression analysis. All three models demonstrated a high segmentation performance, with the UNet3 + model achieving the best performance (DSC 0.95 ± 0.03). Pearson correlation coefficient between the pectoralis and L3 muscle areas showed a significant positive correlation (r = 0.65). The correlation coefficient between the transformed pectoralis and L3 muscle areas showed a stronger positive correlation in both univariate analysis using only muscle area (r = 0.74) and multivariate analysis considering sex, weight, age, and muscle area (r = 0.83). Segmentation of the pectoralis muscle area using artificial intelligence (AI) on chest CT was highly accurate, and the measured values showed a strong correlation with the L3 muscle area. Chest CT using AI technology could play a significant role in the diagnosis of sarcopenia.
Collapse
Affiliation(s)
- Joo Chan Choi
- Department of Biomedical Engineering, College of Health & Science, Gachon University, 191, Hambangmoe-ro, Yeonsu-gu, Incheon, 21936, Republic of Korea
| | - Young Jae Kim
- Gachon Biomedical & Convergence Institute, Gachon University Gil Medical Center, 21, Namdong-daero 774beon-gil, Namdong-gu, Incheon, 21565, Republic of Korea
| | - Kwang Gi Kim
- Department of Biomedical Engineering, College of Health & Science, Gachon University, 191, Hambangmoe-ro, Yeonsu-gu, Incheon, 21936, Republic of Korea.
- Department of Biomedical Engineering, College of Medicine, Gil Medical Center, Gachon University, 38-13 Dokjeom-Ro 3Beon-Gil, Namdong-Gu, Incheon, 21565, Republic of Korea.
- Department of Health Science & Technology, Gachon Advanced Institute for Health Science & Technology (GAIHIST), Lee Gil Ya Cancer and Diabetes Institute, Gachon University, 155 Gaetbeol-Ro, Yeonsu-Gu, Incheon, 21999, Republic of Korea.
| | - Eun Young Kim
- Department of Radiology, Gil Medical Center, Gachon University College of Medicien 21, Namdong-Daero 774Beon-Gil, Namdong-Gu, Incheon, 21565, Republic of Korea.
- Radiology Department, Incheon Sejong Hospital, 20, Gyeyangmunhwa-Ro, Gyeyang-Gu, Incheon, 21080, Republic of Korea.
| |
Collapse
|
10
|
Chen W, Wang H, Zhang L, Zhang M. Temporal and spatial self supervised learning methods for electrocardiograms. Sci Rep 2025; 15:6029. [PMID: 39972080 PMCID: PMC11839927 DOI: 10.1038/s41598-025-90084-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2024] [Accepted: 02/10/2025] [Indexed: 02/21/2025] Open
Abstract
The limited availability of labeled ECG data restricts the application of supervised deep learning methods in ECG detection. Although existing self-supervised learning approaches have been applied to ECG analysis, they are predominantly image-based, which limits their effectiveness. To address these limitations and provide novel insights, we propose a Temporal-Spatial Self-Supervised Learning (TSSL) method specifically designed for ECG detection. TSSL leverages the intrinsic temporal and spatial characteristics of ECG signals to enhance feature representation. Temporally, ECG signals retain consistent identity information over time, enabling the model to generate stable representations for the same individual across different time points while isolating representations of different leads to preserve their unique features. Spatially, ECG signals from various leads capture the heart's activity from different perspectives, revealing both commonalities and distinct patterns. TSSL captures these correlations by maintaining consistency in the relationships between signals and their representations across different leads. Experimental results on the CPSC2018, Chapman, and PTB-XL databases demonstrate that TSSL introduces new capabilities by effectively utilizing temporal and spatial information, achieving superior performance compared to existing methods and approaching the performance of full-label training with only 10% of the labeled data. This highlights TSSL's ability to provide deeper insights and enhanced feature extraction beyond mere performance improvements. We make our code publicly available on https://github.com/cwp9731/temporal-spatial-self-supervised-learning.
Collapse
Affiliation(s)
- Wenping Chen
- College of Information Science and Engineering, Hohai University, Nanjing, 211100, China
| | - Huibin Wang
- College of Computer Science and Software Engineering, Hohai University, Nanjing, 211100, China.
| | - Lili Zhang
- College of Computer Science and Software Engineering, Hohai University, Nanjing, 211100, China
| | - Min Zhang
- College of Information Science and Engineering, Hohai University, Nanjing, 211100, China
- College of Information Engineering, Gannan University of Science and Technology, Ganzhou, 341000, China
| |
Collapse
|
11
|
Wang S, Du G, Dai S, Miao M, Zhang M. Efficient information exchange approach for medical IoT based on AI and DAG-enabled blockchain. Heliyon 2025; 11:e41617. [PMID: 39877624 PMCID: PMC11773007 DOI: 10.1016/j.heliyon.2024.e41617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Revised: 12/28/2024] [Accepted: 12/31/2024] [Indexed: 01/31/2025] Open
Abstract
The development of artificial intelligence (AI) based medical Internet of Things (IoT) technology plays a crucial role in making the collection and exchange of medical information more convenient. However, security, privacy, and efficiency issues during information exchange have become pressing challenges. While many scholars have proposed solutions based on AI and blockchain to address these issues, few have focused on the impact of the slow consensus algorithm of blockchain on the efficiency of information exchange. To improve the efficiency of information exchange, we propose an information exchange approach based on AI and DAG-enabled blockchain, providing a secure and efficient environment for information exchange in the medical IoT. Additionally, to enhance the efficiency of information exchange in the medical IoT, a novel tip selection algorithm is introduced to reduce the time delay in reaching consensus, thereby enabling faster acquisition of trusted information via blockchain. Simulation results demonstrate that compared to methods based on traditional DAG-enabled blockchain, the approach proposed in this paper improves the efficiency of information exchange.
Collapse
Affiliation(s)
- Shanqin Wang
- School of Information Engineering, Chuzhou Polytechnic, Chuzhou, 239000, Anhui, China
| | - Gangxin Du
- Sany Energy Equipment Co., Ltd, Zhuzhou, 412000, Hunan, China
| | - Shufan Dai
- School of Information Engineering, Chuzhou Polytechnic, Chuzhou, 239000, Anhui, China
| | - Mengjun Miao
- School of Information Engineering, Chuzhou Polytechnic, Chuzhou, 239000, Anhui, China
- School of Computer, Qinghai Normal University, Haihu Avenue Chengbei District, Xining, 810008, Qinghai, China
| | - Miao Zhang
- School of Information Engineering, Chuzhou Polytechnic, Chuzhou, 239000, Anhui, China
| |
Collapse
|
12
|
Gadag V, Singh S, Khatri AH, Mishra S, Satapathy SK, Cho SB, Chowdhury A, Pal A, Mohanty SN. Improving myocardial infarction diagnosis with Siamese network-based ECG analysis. PLoS One 2025; 20:e0313390. [PMID: 39883662 PMCID: PMC11781727 DOI: 10.1371/journal.pone.0313390] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Accepted: 10/16/2024] [Indexed: 02/01/2025] Open
Abstract
BACKGROUND Heart muscle damage from myocardial infarction (MI) is brought on by insufficient blood flow. The leading cause of death for middle-aged and older people worldwide is myocardial infarction (MI), which is difficult to diagnose because it has no symptoms. Clinicians must evaluate electrocardiography (ECG) signals to diagnose MI, which is difficult and prone to observer bias. To be effective in actual practice, an automated, and computerized detection system for Myocardial Infarction using ECG images, must meet a number of criteria. OBJECTIVE In an actual clinical situation, these requirements-such as dependability, simplicity, and superior decision-making abilities-remain crucial. In the current work, we have developed a model using a dataset that consists of a combination of 928 ECG images taken from publicly available Mendeley Data. It was converted into three classes Myocardial Infarction, Abnormal heartbeat, and Normal. METHODS The dataset is then imported, pre-processed, and split into a 70:20:10 ratio of training, validation, and testing. It is then trained using the Siamese Network Model. RESULTS The classification accuracy comes out to be 98%. The algorithm works excellently with datasets having class imbalance by taking pair of images as input. The validation and testing classification matrix is then generated and the evaluation metrics for both of them come out to be a near-perfect value. CONCLUSION In this study, we developed the ECG signals based early detection of cardiovascular diseases with Siamese network model.
Collapse
Affiliation(s)
- Vaibhav Gadag
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | - Simrat Singh
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | - Anshul Harish Khatri
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | - Shruti Mishra
- Centre for Advanced Data Science, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | | | - Sung-Bae Cho
- Department of Computer Science, Yonsei University, Seoul, South Korea
| | - Abishi Chowdhury
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | - Amrit Pal
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India
| | - Sachi Nandan Mohanty
- School of Computer Science & Engineering (SCOPE), VIT-AP University, Amaravati, Andhra Pradesh, India
| |
Collapse
|
13
|
Saadati S, Sepahvand A, Razzazi M. Cloud and IoT based smart agent-driven simulation of human gait for detecting muscles disorder. Heliyon 2025; 11:e42119. [PMID: 39906796 PMCID: PMC11791118 DOI: 10.1016/j.heliyon.2025.e42119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Revised: 01/17/2025] [Accepted: 01/18/2025] [Indexed: 02/06/2025] Open
Abstract
Motion disorders affect a significant portion of the global population. While some symptoms can be managed with medications, these treatments often impact all muscles uniformly, not just the affected ones, leading to potential side effects including involuntary movements, confusion, and decreased short-term memory. Currently, there is no dedicated application for differentiating healthy muscles from abnormal ones. Existing analysis applications, designed for other purposes, often lack essential software engineering features such as a user-friendly interface, infrastructure independence, usability and learning ability, cloud computing capabilities, and AI-based assistance. This research proposes a computer-based methodology to analyze human motion and differentiate between healthy and unhealthy muscles. First, an IoT-based approach is proposed to digitize human motion using smartphones instead of hardly accessible wearable sensors and markers. The motion data is then simulated to analyze the neuromusculoskeletal system. An agent-driven modeling method ensures the naturalness, accuracy, and interpretability of the simulation, incorporating neuromuscular details such as Henneman's size principle, action potentials, motor units, and biomechanical principles. The results are then provided to medical and clinical experts to aid in differentiating between healthy and unhealthy muscles and for further investigation. Additionally, a deep learning-based ensemble framework is proposed to assist in the analysis of the simulation results, offering both accuracy and interpretability. A user-friendly graphical interface enhances the application's usability. Being fully cloud-based, the application is infrastructure-independent and can be accessed on smartphones, PCs, and other devices without installation. This strategy not only addresses the current challenges in treating motion disorders but also paves the way for other clinical simulations by considering both scientific and computational requirements.
Collapse
Affiliation(s)
- Sina Saadati
- Department of Computer Engineering, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran
| | - Abdolah Sepahvand
- Department of Computer Engineering, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran
| | - Mohammadreza Razzazi
- Department of Computer Engineering, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran
| |
Collapse
|
14
|
Leone A, Di Napoli V, Fochi NP, Di Perna G, Spetzger U, Filimonova E, Angileri F, Carbone F, Colamaria A. Virtual Biopsy for the Prediction of MGMT Promoter Methylation in Gliomas: A Comprehensive Review of Radiomics and Deep Learning Approaches Applied to MRI. Diagnostics (Basel) 2025; 15:251. [PMID: 39941181 PMCID: PMC11816478 DOI: 10.3390/diagnostics15030251] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2024] [Revised: 01/18/2025] [Accepted: 01/20/2025] [Indexed: 02/16/2025] Open
Abstract
Background/Objectives: The methylation status of the O6-methylguanine-DNA methyltransferase (MGMT) promoter in gliomas has emerged as a critical biomarker for prognosis and treatment response. Conventional methods for assessing MGMT promoter methylation, such as methylation-specific PCR, are invasive and require tissue sampling. Methods: A comprehensive literature search was performed in compliance with the updated PRISMA 2020 guidelines within electronic databases MEDLINE/PubMed, Scopus, and IEEE Xplore. Search terms, including "MGMT", "methylation", "glioma", "glioblastoma", "machine learning", "deep learning", and "radiomics", were adopted in various MeSH combinations. Original studies in the English, Italian, German, and French languages were considered for inclusion. Results: This review analyzed 34 studies conducted in the last six years, focusing on assessing MGMT methylation status using radiomics (RD), deep learning (DL), or combined approaches. These studies utilized radiological data from the public (e.g., BraTS, TCGA) and private institutional datasets. Sixteen studies focused exclusively on glioblastoma (GBM), while others included low- and high-grade gliomas. Twenty-seven studies reported diagnostic accuracy, with fourteen achieving values above 80%. The combined use of DL and RD generally resulted in higher accuracy, sensitivity, and specificity, although some studies reported lower minimum accuracy compared to studies using a single model. Conclusions: The integration of RD and DL offers a powerful, non-invasive tool for precisely recognizing MGMT promoter methylation status in gliomas, paving the way for enhanced personalized medicine in neuro-oncology. The heterogeneity of study populations, data sources, and methodologies reflected the complexity of the pipeline and machine learning algorithms, which may require general standardization to be implemented in clinical practice.
Collapse
Affiliation(s)
- Augusto Leone
- Department of Neurosurgery, Karlsruher Neurozentrum, Städtisches Klinikum Karlsruhe, 76133 Karlsruhe, Germany; (A.L.); (U.S.); (F.C.)
- Faculty of Human Medicine, Charité Universitätsmedizin, 10117 Berlin, Germany
| | - Veronica Di Napoli
- Department of Neurosurgery, University of Turin, 10124 Turin, Italy; (V.D.N.); (N.P.F.)
| | - Nicola Pio Fochi
- Department of Neurosurgery, University of Turin, 10124 Turin, Italy; (V.D.N.); (N.P.F.)
| | - Giuseppe Di Perna
- Division of Neurosurgery, “Policlinico Riuniti”, 71122 Foggia, Italy;
| | - Uwe Spetzger
- Department of Neurosurgery, Karlsruher Neurozentrum, Städtisches Klinikum Karlsruhe, 76133 Karlsruhe, Germany; (A.L.); (U.S.); (F.C.)
| | - Elena Filimonova
- Department of Neuroradiology, Federal Neurosurgical Center, 630048 Novosibirsk, Russia;
| | - Flavio Angileri
- Department of Neurosurgery, University of Messina, 98122 Messina, Italy;
| | - Francesco Carbone
- Department of Neurosurgery, Karlsruher Neurozentrum, Städtisches Klinikum Karlsruhe, 76133 Karlsruhe, Germany; (A.L.); (U.S.); (F.C.)
- Division of Neurosurgery, “Policlinico Riuniti”, 71122 Foggia, Italy;
| | - Antonio Colamaria
- Division of Neurosurgery, “Policlinico Riuniti”, 71122 Foggia, Italy;
| |
Collapse
|
15
|
Rai HM, Yoo J, Agarwal S, Agarwal N. LightweightUNet: Multimodal Deep Learning with GAN-Augmented Imaging Data for Efficient Breast Cancer Detection. Bioengineering (Basel) 2025; 12:73. [PMID: 39851348 PMCID: PMC11761908 DOI: 10.3390/bioengineering12010073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2024] [Revised: 01/06/2025] [Accepted: 01/08/2025] [Indexed: 01/26/2025] Open
Abstract
Breast cancer ranks as the second most prevalent cancer globally and is the most frequently diagnosed cancer among women; therefore, early, automated, and precise detection is essential. Most AI-based techniques for breast cancer detection are complex and have high computational costs. Hence, to overcome this challenge, we have presented the innovative LightweightUNet hybrid deep learning (DL) classifier for the accurate classification of breast cancer. The proposed model boasts a low computational cost due to its smaller number of layers in its architecture, and its adaptive nature stems from its use of depth-wise separable convolution. We have employed a multimodal approach to validate the model's performance, using 13,000 images from two distinct modalities: mammogram imaging (MGI) and ultrasound imaging (USI). We collected the multimodal imaging datasets from seven different sources, including the benchmark datasets DDSM, MIAS, INbreast, BrEaST, BUSI, Thammasat, and HMSS. Since the datasets are from various sources, we have resized them to the uniform size of 256 × 256 pixels and normalized them using the Box-Cox transformation technique. Since the USI dataset is smaller, we have applied the StyleGAN3 model to generate 10,000 synthetic ultrasound images. In this work, we have performed two separate experiments: the first on a real dataset without augmentation and the second on a real + GAN-augmented dataset using our proposed method. During the experiments, we used a 5-fold cross-validation method, and our proposed model obtained good results on the real dataset (87.16% precision, 86.87% recall, 86.84% F1-score, and 86.87% accuracy) without adding any extra data. Similarly, the second experiment provides better performance on the real + GAN-augmented dataset (96.36% precision, 96.35% recall, 96.35% F1-score, and 96.35% accuracy). This multimodal approach, which utilizes LightweightUNet, enhances the performance by 9.20% in precision, 9.48% in recall, 9.51% in F1-score, and a 9.48% increase in accuracy on the combined dataset. The LightweightUNet model we proposed works very well thanks to a creative network design, adding fake images to the data, and a multimodal training method. These results show that the model has a lot of potential for use in clinical settings.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea;
| | - Joon Yoo
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea;
| | - Saurabh Agarwal
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
| | - Neha Agarwal
- School of Chemical Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
| |
Collapse
|
16
|
Motallebi S, Zandieh M, Tabriz AA, Tirkolaee EB. Assessing the industry 4.0 strategies for a steel supply chain: SWOT, game theory, and gap analysis. Heliyon 2025; 11:e41374. [PMID: 39811319 PMCID: PMC11730224 DOI: 10.1016/j.heliyon.2024.e41374] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2024] [Revised: 12/07/2024] [Accepted: 12/18/2024] [Indexed: 01/16/2025] Open
Abstract
The recent adoption of modern technologies has led to the fourth industrial revolution or Industry 4.0 (I4.0). The supply chain of large industries, such as steel, is of significant importance in the economies of countries. According to data published by the National Iranian Steel Company, the steel industry is one of the most vital sectors in Iran, contributing about 1 % to its Gross Domestic Product (GDP). Therefore, implementing I4.0 in the steel supply chain is a highly significant and practical endeavor, with the potential to drive considerable advancements in this industry. This study develops an efficient integrated framework to identify and evaluate the most critical factors influencing the implementation of I4.0 in the supply chains of large industries. The study also reviews the most effective policies for implementing these factors and analyzes the gap between the current state of the steel supply chain in Iran and the desired state within the framework of I4.0. The components of I4.0 are determined based on a review of existing literature and interviews with experts from the Iranian steel industry, utilizing the fuzzy Delphi method. Afterward, Strengths, Weaknesses, Opportunities, and Threats (SWOT) policies are developed. The optimal combination of policies for implementing I4.0 in the Iranian steel supply chain is identified using game theory. Finally, a gap analysis is conducted between the current state of the industry and the desired state. The critical gap analysis revealed that the current state of the supply chain significantly lags behind the desired I4.0 goals, particularly in terms of infrastructure and supply chain balancing. The findings highlight a stark contrast between the current performance score of 0.16 and the target score of 0.84, underscoring the substantial improvements needed to realize I4.0 within this sector.
Collapse
Affiliation(s)
- Sima Motallebi
- Department of Industrial Management and Information Technology, Faculty of Management and Accounting, Shahid Beheshti University, Tehran, Iran
| | - Mostafa Zandieh
- Department of Industrial Management and Information Technology, Faculty of Management and Accounting, Shahid Beheshti University, Tehran, Iran
| | - Akbar Alem Tabriz
- Department of Industrial Management and Information Technology, Faculty of Management and Accounting, Shahid Beheshti University, Tehran, Iran
| | - Erfan Babaee Tirkolaee
- Department of Industrial Engineering, Istinye University, Istanbul, Turkey
- Department of Industrial Engineering and Management, Yuan Ze University, Taoyuan, Taiwan
- Department of Mechanics and Mathematics, Western Caspian University, Baku, Azerbaijan
| |
Collapse
|
17
|
Rai HM, Yoo J, Dashkevych S. Transformative Advances in AI for Precise Cancer Detection: A Comprehensive Review of Non-Invasive Techniques. ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING 2025. [DOI: 10.1007/s11831-024-10219-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Accepted: 12/07/2024] [Indexed: 03/02/2025]
|
18
|
Rastogi D, Johri P, Donelli M, Kadry S, Khan AA, Espa G, Feraco P, Kim J. Deep learning-integrated MRI brain tumor analysis: feature extraction, segmentation, and Survival Prediction using Replicator and volumetric networks. Sci Rep 2025; 15:1437. [PMID: 39789043 PMCID: PMC11718254 DOI: 10.1038/s41598-024-84386-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Accepted: 12/23/2024] [Indexed: 01/12/2025] Open
Abstract
The most prevalent form of malignant tumors that originate in the brain are known as gliomas. In order to diagnose, treat, and identify risk factors, it is crucial to have precise and resilient segmentation of the tumors, along with an estimation of the patients' overall survival rate. Therefore, we have introduced a deep learning approach that employs a combination of MRI scans to accurately segment brain tumors and predict survival in patients with gliomas. To ensure strong and reliable tumor segmentation, we employ 2D volumetric convolution neural network architectures that utilize a majority rule. This method helps to significantly decrease model bias and improve performance. Additionally, in order to predict survival rates, we extract radiomic features from the tumor regions that have been segmented, and then use a Deep Learning Inspired 3D replicator neural network to identify the most effective features. The model presented in this study was successful in segmenting brain tumors and predicting the outcome of enhancing tumor and real enhancing tumor. The model was evaluated using the BRATS2020 benchmarks dataset, and the obtained results are quite satisfactory and promising.
Collapse
Affiliation(s)
- Deependra Rastogi
- School of Computer Science and Engineering, IILM University, Greater Noida, Noida, 201306, UP, India
| | - Prashant Johri
- SCSE, Galgotias University, Greater Noida, Noida, 203201, UP, India
| | - Massimo Donelli
- Department of Civil, Environmental, Mechanical Engineering University of Trento, Trento, 38100, Italy
- Radiomics Laboratory, Department of Economy and Management, University of Trento, Trento, 38100, Italy
| | - Seifedine Kadry
- Department of Computer Science and Mathematics, Lebanese American University, Beirut, Lebanon
- Noroff University College, Kristiansand, 4612, Norway
| | - Arfat Ahmad Khan
- Department of Engineering, Simpson University, California, 96003, USA.
| | - Giuseppe Espa
- Radiomics Laboratory, Department of Economy and Management, University of Trento, Trento, 38100, Italy
| | - Paola Feraco
- Neuroradiology Unit, Santa Chiara Hospital, Azienda provinciale per i servizi sanitari, Trento, 38100, Italy
| | - Jungeun Kim
- Department of Computer Engineering, Inha University, Incheon, Republic of Korea.
| |
Collapse
|
19
|
Lee S, Shin C, Kang HG, Lee S. Recurrent Flow Update Model Using Image Pyramid Structure for 4K Video Frame Interpolation. SENSORS (BASEL, SWITZERLAND) 2025; 25:290. [PMID: 39797081 PMCID: PMC11723460 DOI: 10.3390/s25010290] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/14/2024] [Revised: 01/02/2025] [Accepted: 01/05/2025] [Indexed: 01/13/2025]
Abstract
Video frame interpolation (VFI) is a task that generates intermediate frames from two consecutive frames. Previous studies have employed two main approaches to extract the necessary information from both frames: pixel-level synthesis and flow-based methods. However, when synthesizing high-resolution videos using VFI, each approach has its limitations. Pixel-level synthesis based on the transformer architecture requires high complexity to achieve 4K video results. In the case of flow-based methods, forward warping can produce holes where pixels are not allocated, while backward warping approaches struggle to obtain accurate backward flow. Additionally, there are challenges during the training stage; previous works have often generated suboptimal results by training multi-stage model architectures separately. To address these issues, we propose a Recurrent Flow Update (RFU) model trained in an end-to-end manner. We introduce a global flow update module that leverages global information to mitigate the weaknesses of forward flow and gradually correct errors. We demonstrate the effectiveness of our method through several ablation studies. Our approach achieves state-of-the-art performance not only on the XTest and Davis datasets, which have 4K resolution, but also on the SNU-FILM dataset, which features large motions at low resolution.
Collapse
Affiliation(s)
| | | | | | - Sangyoun Lee
- School of Electrical and Electronic Engineering, Yonsei University, Seoul 03722, Republic of Korea; (S.L.); (C.S.); (H.-G.K.)
| |
Collapse
|
20
|
Eidler P, Kopylov U, Ukashi O. Capsule Endoscopy in Inflammatory Bowel Disease: Evolving Role and Recent Advances. Gastrointest Endosc Clin N Am 2025; 35:73-102. [PMID: 39510694 DOI: 10.1016/j.giec.2024.07.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2024]
Abstract
Capsule endoscopy has been proven as an efficient and accurate tool in the diagnosing and monitoring patients with inflammatory bowel disease, especially Crohn's disease (CD). The current European Crohn's and Colitis Organization guidelines recommend small bowel disease assessment in newly diagnosed CD, wherein small bowel capsule endoscopy (SBCE) is of prime importance. SBCE plays an essential role in assessing mucosal healing in patients with CD, serving as a monitoring tool in a treat to target strategy, and is capable of identifying high-risk patients for future flares.
Collapse
Affiliation(s)
- Pinhas Eidler
- Gastroenterology Institute, Sheba Medical Center Tel Hashomer, Ramat Gan 52621, Israel
| | - Uri Kopylov
- Gastroenterology Institute, Sheba Medical Center Tel Hashomer, Ramat Gan 52621, Israel; Faculty of Medical and Health Sciences, Tel-Aviv University, Ramat Aviv, Tel Aviv 69978, Israel
| | - Offir Ukashi
- Gastroenterology Institute, Sheba Medical Center Tel Hashomer, Ramat Gan 52621, Israel; Faculty of Medical and Health Sciences, Tel-Aviv University, Ramat Aviv, Tel Aviv 69978, Israel.
| |
Collapse
|
21
|
Hu Q, Wang D, Wu H, Liu J, Yang C. Efficient multi-view fusion and flexible adaptation to view missing in cardiovascular system signals. Neural Netw 2025; 181:106760. [PMID: 39362184 DOI: 10.1016/j.neunet.2024.106760] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2023] [Revised: 09/19/2024] [Accepted: 09/23/2024] [Indexed: 10/05/2024]
Abstract
The progression of deep learning and the widespread adoption of sensors have facilitated automatic multi-view fusion (MVF) about the cardiovascular system (CVS) signals. However, prevalent MVF model architecture often amalgamates CVS signals from the same temporal step but different views into a unified representation, disregarding the asynchronous nature of cardiovascular events and the inherent heterogeneity across views, leading to catastrophic view confusion. Efficient training strategies specifically tailored for MVF models to attain comprehensive representations need simultaneous consideration. Crucially, real-world data frequently arrives with incomplete views, an aspect rarely noticed by researchers. Thus, the View-Centric Transformer (VCT) and Multitask Masked Autoencoder (M2AE) are specifically designed to emphasize the centrality of each view and harness unlabeled data to achieve superior fused representations. Additionally, we systematically define the missing-view problem for the first time and introduce prompt techniques to aid pretrained MVF models in flexibly adapting to various missing-view scenarios. Rigorous experiments involving atrial fibrillation detection, blood pressure estimation, and sleep staging-typical health monitoring tasks-demonstrate the remarkable advantage of our method in MVF compared to prevailing methodologies. Notably, the prompt technique requires finetuning <3 % of the entire model's data, substantially fortifying the model's resilience to view missing while circumventing the need for complete retraining. The results demonstrate the effectiveness of our approaches, highlighting their potential for practical applications in cardiovascular health monitoring. Codes and models are released at URL.
Collapse
Affiliation(s)
- Qihan Hu
- Center for Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai, 200433, PR China
| | - Daomiao Wang
- Center for Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai, 200433, PR China
| | - Hong Wu
- Center for Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai, 200433, PR China
| | - Jian Liu
- Center for Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai, 200433, PR China
| | - Cuiwei Yang
- Center for Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai, 200433, PR China; Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention of Shanghai, 200093, PR China.
| |
Collapse
|
22
|
Wang X, Yue X, Sajid A, Tariq N. AllianceBlockchain in the Governance Innovation of Internet Hospitals. SENSORS (BASEL, SWITZERLAND) 2024; 25:142. [PMID: 39796933 PMCID: PMC11723095 DOI: 10.3390/s25010142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/21/2024] [Revised: 12/25/2024] [Accepted: 12/26/2024] [Indexed: 01/13/2025]
Abstract
The rise of Internet hospitals has significant issues associated with data security and governance in managing sensitive patient data. This paper discusses an alliance blockchain (i.e., a private blockchain) model for governance innovation in internet hospitals with an improved encryption methodology. We compare our proposed model, improved Rivest-Shamir-Adleman (RSA) encryption, integrated into the blockchain framework. Improved RSA achieves impressive improvements in all key metrics by increasing the throughput by 24.7% and lowering the latency by 19.8% compared to the base model. Thus, the improved model is more optimized for processing transactions related to healthcare data. Memory usage was also reduced by 14.3%. While encryption time remained pretty close, the decryption time remarkably improved by 97.5%. IoT sensors are one of the foundations for Internet hospitals that produce consistent patient data streams, such as physiological and environmental metrics. The proposed alliance blockchain model enables the secure and efficient real-time management of this sensor data. These results demonstrate the capability of alliance blockchain and cryptographic upgrades in creating safe and efficient governance frameworks for Internet hospitals.
Collapse
Affiliation(s)
- Xiaofeng Wang
- College of Management, Shenzhen University, Shenzhen 518060, China
| | - Xiaoguang Yue
- College of Management, Shenzhen University, Shenzhen 518060, China
| | - Ahthasham Sajid
- Multimedia University, Cyberjaya 63100, Malaysia;
- Department of Information Security and Data Science, Riphah Institute of Systems Engineering, Riphah International University, Islamabad 46000, Pakistan
| | - Noshina Tariq
- Department of Artificial Intelligence and Data Science, National University of Computer and Emerging Sciences, Islamabad 44000, Pakistan;
| |
Collapse
|
23
|
Rai HM, Yoo J, Agarwal S. The Improved Network Intrusion Detection Techniques Using the Feature Engineering Approach with Boosting Classifiers. MATHEMATICS 2024; 12:3909. [DOI: 10.3390/math12243909] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2025]
Abstract
In the domain of cybersecurity, cyber threats targeting network devices are very crucial. Because of the exponential growth of wireless devices, such as smartphones and portable devices, cyber risks are becoming increasingly frequent and common with the emergence of new types of threats. This makes the automatic and accurate detection of network-based intrusion very essential. In this work, we propose a network-based intrusion detection system utilizing the comprehensive feature engineering approach combined with boosting machine-learning (ML) models. A TCP/IP-based dataset with 25,192 data samples from different protocols has been utilized in our work. To improve the dataset, we used preprocessing methods such as label encoding, correlation analysis, custom label encoding, and iterative label encoding. To improve the model’s accuracy for prediction, we then used a unique feature engineering methodology that included novel feature scaling and random forest-based feature selection techniques. We used three conventional models (NB, LR, and SVC) and four boosting classifiers (CatBoostGBM, LightGBM, HistGradientBoosting, and XGBoost) for classification. The 10-fold cross-validation methods were employed to train each model. After an assessment using numerous metrics, the best-performing model emerged as XGBoost. With mean metric values of 99.54 ± 0.0007 for accuracy, 99.53 ± 0.0013 for precision, 99.54 ± 0.001 for recall, and an F1-score of 99.53 ± 0.0014, the XGBoost model produced the best performance overall. Additionally, we showed the ROC curve for evaluating the model, which demonstrated that all boosting classifiers obtained a perfect AUC value of one. Our suggested methodologies show effectiveness and accuracy in detecting network intrusions, setting the stage for the model to be used in real time. Our method provides a strong defensive measure against malicious intrusions into network infrastructures while cyber threats keep varying.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam 13120, Republic of Korea
| | - Joon Yoo
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam 13120, Republic of Korea
| | - Saurabh Agarwal
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
| |
Collapse
|
24
|
Singh DD, Haque S, Kim Y, Han I, Yadav DK. Remodeling of tumour microenvironment: strategies to overcome therapeutic resistance and innovate immunoengineering in triple-negative breast cancer. Front Immunol 2024; 15:1455211. [PMID: 39720730 PMCID: PMC11666570 DOI: 10.3389/fimmu.2024.1455211] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Accepted: 10/31/2024] [Indexed: 12/26/2024] Open
Abstract
Triple-negative breast cancer (TNBC) stands as the most complex and daunting subtype of breast cancer affecting women globally. Regrettably, treatment options for TNBC remain limited due to its clinical complexity. However, immunotherapy has emerged as a promising avenue, showing success in developing effective therapies for advanced cases and improving patient outcomes. Improving TNBC treatments involves reducing side effects, minimizing systemic toxicity, and enhancing efficacy. Unlike traditional cancer immunotherapy, engineered nonmaterial's can precisely target TNBC, facilitating immune cell access, improving antigen presentation, and triggering lasting immune responses. Nanocarriers with enhanced sensitivity and specificity, specific cellular absorption, and low toxicity are gaining attention. Nanotechnology-driven immunoengineering strategies focus on targeted delivery systems using multifunctional molecules for precise tracking, diagnosis, and therapy in TNBC. This study delves into TNBC's tumour microenvironment (TME) remodeling, therapeutic resistance, and immunoengineering strategies using nanotechnology.
Collapse
Affiliation(s)
- Desh Deepak Singh
- Amity Institute of Biotechnology, Amity University Rajasthan, Jaipur, India
| | - Shafiul Haque
- Research and Scientific Studies Unit, College of Nursing and Allied Health Sciences, Jazan University, Jazan, Saudi Arabia
| | - Youngsun Kim
- Department of Obstetrics and Gynecology, Kyung Hee University Medical Center, Seoul, Republic of Korea
| | - Ihn Han
- Plasma Bioscience Research Center, Applied Plasma Medicine Center, Department of Electrical & Biological Physics, Kwangwoon University, Seoul, Republic of Korea
| | - Dharmendra Kumar Yadav
- Department of Biologics, College of Pharmacy, Hambakmoeiro 191, Yeonsu-gu, Incheon, Republic of Korea
| |
Collapse
|
25
|
Ahmad I, Alqurashi F. Early cancer detection using deep learning and medical imaging: A survey. Crit Rev Oncol Hematol 2024; 204:104528. [PMID: 39413940 DOI: 10.1016/j.critrevonc.2024.104528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Accepted: 10/02/2024] [Indexed: 10/18/2024] Open
Abstract
Cancer, characterized by the uncontrolled division of abnormal cells that harm body tissues, necessitates early detection for effective treatment. Medical imaging is crucial for identifying various cancers, yet its manual interpretation by radiologists is often subjective, labour-intensive, and time-consuming. Consequently, there is a critical need for an automated decision-making process to enhance cancer detection and diagnosis. Previously, a lot of work was done on surveys of different cancer detection methods, and most of them were focused on specific cancers and limited techniques. This study presents a comprehensive survey of cancer detection methods. It entails a review of 99 research articles collected from the Web of Science, IEEE, and Scopus databases, published between 2020 and 2024. The scope of the study encompasses 12 types of cancer, including breast, cervical, ovarian, prostate, esophageal, liver, pancreatic, colon, lung, oral, brain, and skin cancers. This study discusses different cancer detection techniques, including medical imaging data, image preprocessing, segmentation, feature extraction, deep learning and transfer learning methods, and evaluation metrics. Eventually, we summarised the datasets and techniques with research challenges and limitations. Finally, we provide future directions for enhancing cancer detection techniques.
Collapse
Affiliation(s)
- Istiak Ahmad
- Department of Computer Science, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia; School of Information and Communication Technology, Griffith University, Queensland 4111, Australia.
| | - Fahad Alqurashi
- Department of Computer Science, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| |
Collapse
|
26
|
Rai HM, Yoo J, Razaque A. Comparative analysis of machine learning and deep learning models for improved cancer detection: A comprehensive review of recent advancements in diagnostic techniques. EXPERT SYSTEMS WITH APPLICATIONS 2024; 255:124838. [DOI: 10.1016/j.eswa.2024.124838] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/11/2024]
|
27
|
Rai HM, Yoo J, Razaque A. A depth analysis of recent innovations in non-invasive techniques using artificial intelligence approach for cancer prediction. Med Biol Eng Comput 2024; 62:3555-3580. [PMID: 39012415 DOI: 10.1007/s11517-024-03158-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Accepted: 06/22/2024] [Indexed: 07/17/2024]
Abstract
The fight against cancer, a relentless global health crisis, emphasizes the urgency for efficient and automated early detection methods. To address this critical need, this review assesses recent advances in non-invasive cancer prediction techniques, comparing conventional machine learning (CML) and deep neural networks (DNNs). Focusing on these seven major cancers, we analyze 310 publications spanning the years 2018 to 2024, focusing on detection accuracy as the key metric to identify the most effective predictive models, highlighting critical gaps in current methodologies, and suggesting directions for future research. We further delved into factors like datasets, features, and modalities to gain a comprehensive understanding of each approach's performance. Separate review tables for each cancer type and approach facilitated comparisons between top performers (accuracy exceeding 99%) and low performers (65.83 to 85.8%). Our exploration of public databases and commonly used classifiers revealed that optimal combinations of features, datasets, and models can achieve up to 100% accuracy for both CML and DNN. However, significant variations in accuracy (up to 35%) were observed, particularly when optimization was lacking. Notably, colorectal cancer exhibited the lowest accuracy (DNN 69%, CML 65.83%). A five-point comparative analysis (best/worst models, performance gap, average accuracy, and research trends) revealed that while DNN research is gaining momentum, CML approaches remain competitive, even outperforming DNN in some cases. This study presents an in-depth comparative analysis of CML and DNN techniques for cancer detection. This knowledge can inform future research directions and contribute to the development of increasingly accurate and reliable cancer detection tools.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-Gu, Seongnam-Si, 13120, Gyeonggi-Do, Republic of Korea.
| | - Joon Yoo
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-Gu, Seongnam-Si, 13120, Gyeonggi-Do, Republic of Korea
| | - Abdul Razaque
- Department of Cyber Security, Information Processing and Storage, Satbayev University, Almaty, Kazakhstan
| |
Collapse
|
28
|
Pal A, Rai HM, Frej MBH, Razaque A. Advanced Segmentation of Gastrointestinal (GI) Cancer Disease Using a Novel U-MaskNet Model. Life (Basel) 2024; 14:1488. [PMID: 39598286 PMCID: PMC11595444 DOI: 10.3390/life14111488] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2024] [Revised: 11/05/2024] [Accepted: 11/12/2024] [Indexed: 11/29/2024] Open
Abstract
The purpose of this research is to contribute to the development of approaches for the classification and segmentation of various gastrointestinal (GI) cancer diseases, such as dyed lifted polyps, dyed resection margins, esophagitis, normal cecum, normal pylorus, normal Z line, polyps, and ulcerative colitis. This research is relevant and essential because of the current challenges related to the absence of efficient diagnostic tools for early diagnostics of GI cancers, which are fundamental for improving the diagnosis of these common diseases. To address the above challenges, we propose a new hybrid segmentation model, U-MaskNet, which is a combination of U-Net and Mask R-CNN models. Here, U-Net is utilized for pixel-wise classification and Mask R-CNN for instance segmentation, together forming a solution for classifying and segmenting GI cancer. The Kvasir dataset, which includes 8000 endoscopic images of various GI cancers, is utilized to validate the proposed methodology. The experimental results clearly demonstrated that the novel proposed model provided superior segmentation compared to other well-known models, such as DeepLabv3+, FCN, and DeepMask, as well as improved classification performance compared to state-of-the-art (SOTA) models, including LeNet-5, AlexNet, VGG-16, ResNet-50, and the Inception Network. The quantitative analysis revealed that our proposed model outperformed the other models, achieving a precision of 98.85%, recall of 98.49%, and F1 score of 98.68%. Additionally, the novel model achieved a Dice coefficient of 94.35% and IoU of 89.31%. Consequently, the developed model increased the accuracy and reliability in detecting and segmenting GI cancer, and it was proven that the proposed model can potentially be used for improving the diagnostic process and, consequently, patient care in the clinical environment. This work highlights the benefits of integrating the U-Net and Mask R-CNN models, opening the way for further research in medical image segmentation.
Collapse
Affiliation(s)
- Aditya Pal
- Department of Information Technology, Dronacharya Group of Institutions, Greater Noida 201306, India;
| | - Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Gyeonggi-do, Republic of Korea
| | - Mohamed Ben Haj Frej
- Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT 06604, USA
| | - Abdul Razaque
- Department of Electrical, Computer Engineering and Computer Science, Ohio Northern University, Ada, OH 45810, USA
| |
Collapse
|
29
|
Hussain MA, LaMay D, Grant E, Ou Y. Deep learning of structural MRI predicts fluid, crystallized, and general intelligence. Sci Rep 2024; 14:27935. [PMID: 39537706 PMCID: PMC11561325 DOI: 10.1038/s41598-024-78157-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2024] [Accepted: 10/29/2024] [Indexed: 11/16/2024] Open
Abstract
Can brain structure predict human intelligence? T1-weighted structural brain magnetic resonance images (sMRI) have been correlated with intelligence. However, the population-level association does not fully account for individual variability in intelligence. To address this, studies have emerged recently to predict individual subject's intelligence or neurocognitive scores. However, they are mostly on predicting fluid intelligence (the ability to solve new problems). Studies are lacking to predict crystallized intelligence (the ability to accumulate knowledge) or general intelligence (fluid and crystallized intelligence combined). This study tests whether deep learning of sMRI can predict an individual subject's verbal, comprehensive, and full-scale intelligence quotients (VIQ, PIQ, and FSIQ), which reflect fluid and crystallized intelligence. We performed a comprehensive set of 432 experiments, using different input image channels, six deep learning models, and two outcome settings, in 850 healthy and autistic subjects 6-64 years of age. Our findings indicate a statistically significant potential of T1-weighted sMRI in predicting intelligence, with a Pearson correlation exceeding 0.21 (p < 0.001). Interestingly, we observed that an increase in the complexity of deep learning models does not necessarily translate to higher accuracy in intelligence prediction. The interpretations of our 2D and 3D CNNs, based on GradCAM, align well with the Parieto-Frontal Integration Theory (P-FIT), reinforcing the theory's suggestion that human intelligence is a result of interactions among various brain regions, including the occipital, temporal, parietal, and frontal lobes. These promising results invite further studies and open new questions in the field.
Collapse
Affiliation(s)
- Mohammad Arafat Hussain
- Department of Pediatrics, Boston Children's Hospital, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA
| | - Danielle LaMay
- Department of Pediatrics, Boston Children's Hospital, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA
- Khoury College of Computer and Information Science, Northeastern University, 360 Huntington Ave, Boston, MA, 02115, USA
| | - Ellen Grant
- Department of Pediatrics, Boston Children's Hospital, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA
- Department of Radiology, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA
| | - Yangming Ou
- Department of Pediatrics, Boston Children's Hospital, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA.
- Department of Radiology, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA.
- Computational Health Informatics Program, Boston Children's Hospital, Harvard Medical School, 401 Park Drive, Boston, MA, 02115, USA.
| |
Collapse
|
30
|
Kim S, Park H, Park SH. A review of deep learning-based reconstruction methods for accelerated MRI using spatiotemporal and multi-contrast redundancies. Biomed Eng Lett 2024; 14:1221-1242. [PMID: 39465106 PMCID: PMC11502678 DOI: 10.1007/s13534-024-00425-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2024] [Revised: 08/27/2024] [Accepted: 09/06/2024] [Indexed: 10/29/2024] Open
Abstract
Accelerated magnetic resonance imaging (MRI) has played an essential role in reducing data acquisition time for MRI. Acceleration can be achieved by acquiring fewer data points in k-space, which results in various artifacts in the image domain. Conventional reconstruction methods have resolved the artifacts by utilizing multi-coil information, but with limited robustness. Recently, numerous deep learning-based reconstruction methods have been developed, enabling outstanding reconstruction performances with higher acceleration. Advances in hardware and developments of specialized network architectures have produced such achievements. Besides, MRI signals contain various redundant information including multi-coil redundancy, multi-contrast redundancy, and spatiotemporal redundancy. Utilization of the redundant information combined with deep learning approaches allow not only higher acceleration, but also well-preserved details in the reconstructed images. Consequently, this review paper introduces the basic concepts of deep learning and conventional accelerated MRI reconstruction methods, followed by review of recent deep learning-based reconstruction methods that exploit various redundancies. Lastly, the paper concludes by discussing the challenges, limitations, and potential directions of future developments.
Collapse
Affiliation(s)
- Seonghyuk Kim
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| | - HyunWook Park
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| | - Sung-Hong Park
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
- Department of Bio and Brain Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Yuseong-gu, Daejeon, 34141 Republic of Korea
| |
Collapse
|
31
|
Mukherjee J, Sharma R, Dutta P, Bhunia B. Artificial intelligence in healthcare: a mastery. Biotechnol Genet Eng Rev 2024; 40:1659-1708. [PMID: 37013913 DOI: 10.1080/02648725.2023.2196476] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2023] [Accepted: 03/22/2023] [Indexed: 04/05/2023]
Abstract
There is a vast development of artificial intelligence (AI) in recent years. Computational technology, digitized data collection and enormous advancement in this field have allowed AI applications to penetrate the core human area of specialization. In this review article, we describe current progress achieved in the AI field highlighting constraints on smooth development in the field of medical AI sector, with discussion of its implementation in healthcare from a commercial, regulatory and sociological standpoint. Utilizing sizable multidimensional biological datasets that contain individual heterogeneity in genomes, functionality and milieu, precision medicine strives to create and optimize approaches for diagnosis, treatment methods and assessment. With the arise of complexity and expansion of data in the health-care industry, AI can be applied more frequently. The main application categories include indications for diagnosis and therapy, patient involvement and commitment and administrative tasks. There has recently been a sharp rise in interest in medical AI applications due to developments in AI software and technology, particularly in deep learning algorithms and in artificial neural network (ANN). In this overview, we enlisted the major categories of issues that AI systems are ideally equipped to resolve followed by clinical diagnostic tasks. It also includes a discussion of the future potential of AI, particularly for risk prediction in complex diseases, and the difficulties, constraints and biases that must be meticulously addressed for the effective delivery of AI in the health-care sector.
Collapse
Affiliation(s)
- Jayanti Mukherjee
- Department of Pharmaceutical Chemistry, CMR College of Pharmacy Affiliated to Jawaharlal Nehru Technological University, Hyderabad, Telangana, India
| | - Ramesh Sharma
- Department of Bioengineering, National Institute of Technology, Agartala, India
| | - Prasenjit Dutta
- Department of Production Engineering, National Institute of Technology, Agartala, India
| | - Biswanath Bhunia
- Department of Bioengineering, National Institute of Technology, Agartala, India
| |
Collapse
|
32
|
Rai HM, Shukla KK, Tightiz L, Padmanaban S. Enhancing data security and privacy in energy applications: Integrating IoT and blockchain technologies. Heliyon 2024; 10:e38917. [PMID: 39430499 PMCID: PMC11490785 DOI: 10.1016/j.heliyon.2024.e38917] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2024] [Revised: 09/30/2024] [Accepted: 10/02/2024] [Indexed: 10/22/2024] Open
Abstract
The integration of blockchain technology with the IoToffers numerous opportunities to enhance the privacy, security, and integrity. This study comprehensively analyze the challenges, scope, and potential solutions associated with integrating blockchain technology and the IoT, with a specific emphasis on nuclear energy applications. We discuss the roles and various aspects of blockchain and the IoT, highlighting their multiple dimensions and applications. Our study develops a secure data management framework that incorporates encryption, integrity verification, an integrated communication network, and a robust data flow architecture. We explore the several aspects of data security, privacy, and integrity, along with the potential solutions in the integration of blockchain and IoT. The study also investigates the secure transaction process, with a specific focus on cryptographic, mathematical, and algorithmic perspectives. We demonstrated the use of blockchain technology in the nuclear energy sector using flow charts, comprehensively addressing the associated security and privacy concerns. While emphasizing the applicability of our methodology to the nuclear sector, we also acknowledge limitations such as requirements for practical validation, challenges with resource-constrained IoT environments, increasing cyberthreats, and limited real-time data availability. The future scope of our study focuses on standardization, scalable blockchain, post-quantum cryptography, privacy, regulations, real-world testbeds, and deep learning for nuclear sector security. Our findings highlight that the integration of blockchain and IoT can significantly enhance the security and privacy of nuclear energy applications, although practical validation and optimization are necessary.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si, 13120, Gyeonggi-do, Republic of Korea
| | | | - Lilia Tightiz
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si, 13120, Gyeonggi-do, Republic of Korea
| | - Sanjeevikumar Padmanaban
- Department of Electrical Engineering, IT and Cybernetics, University of South-Eastern Norway, Porsgrunn, 3918, Norway
| |
Collapse
|
33
|
Ghantasala GSP, Dilip K, Vidyullatha P, Allabun S, Alqahtani MS, Othman M, Abbas M, Soufiene BO. Enhanced ovarian cancer survival prediction using temporal analysis and graph neural networks. BMC Med Inform Decis Mak 2024; 24:299. [PMID: 39390514 PMCID: PMC11468212 DOI: 10.1186/s12911-024-02665-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2024] [Accepted: 09/04/2024] [Indexed: 10/12/2024] Open
Abstract
Ovarian cancer is a formidable health challenge that demands accurate and timely survival predictions to guide clinical interventions. Existing methods, while commendable, suffer from limitations in harnessing the temporal evolution of patient data and capturing intricate interdependencies among different data elements. In this paper, we present a novel methodology which combines Temporal Analysis and Graph Neural Networks (GNNs) to significantly enhance ovarian cancer survival rate predictions. The shortcomings of current processes originate from their disability to correctly seize the complex interactions amongst diverse scientific information units in addition to the dynamic modifications that arise in a affected person`s nation over time. By combining temporal information evaluation and GNNs, our cautioned approach overcomes those drawbacks and, whilst as compared to preceding methods, yields a noteworthy 8.3% benefit in precision, 4.9% more accuracy, 5.5% more advantageous recall, and a considerable 2.9% reduction in prediction latency. Our method's Temporal Analysis factor uses longitudinal affected person information to perceive good-sized styles and tendencies that offer precious insights into the direction of ovarian cancer. Through the combination of GNNs, we offer a robust framework able to shoot complicated interactions among exclusive capabilities of scientific data, permitting the version to realize diffused dependencies that would affect survival results. Our paintings have tremendous implications for scientific practice. Prompt and correct estimation of the survival price of ovarian most cancers allows scientific experts to customize remedy regimens, manipulate assets efficiently, and provide individualized care to patients. Additionally, the interpretability of our version`s predictions promotes a collaborative method for affected person care via way of means of strengthening agreement among scientific employees and the AI-driven selection help system. The proposed approach not only outperforms existing methods but also has the possible to develop ovarian cancer treatment by providing clinicians through a reliable tool for informed decision-making. Through a fusion of Temporal Analysis and Graph Neural Networks, we conduit the gap among data-driven insights and clinical practice, proposing a capable opportunity for refining patient outcomes in ovarian cancer management operations.
Collapse
Affiliation(s)
| | - Kumar Dilip
- Department of Computer Science and Engineering, Alliance University, Bengaluru, India
| | - Pellakuri Vidyullatha
- Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Guntur, Andhra Pradesh, India
| | - Sarah Allabun
- Department of Medical Education, College of Medicine, Princess Nourah bint Abdulrahman University, P.O.Box 84428, Riyadh, 11671, Saudi Arabia
| | - Mohammed S Alqahtani
- Radiological Sciences Department, College of Applied Medical Sciences, King Khalid University, Abha, 61421, Saudi Arabia
- BioImaging Unit, Space Research Centre, University of Leicester, Michael Atiyah Building, Leicester, LE1 7RH, UK
| | - Manal Othman
- Department of Medical Education, College of Medicine, Princess Nourah bint Abdulrahman University, P.O.Box 84428, Riyadh, 11671, Saudi Arabia
| | - Mohamed Abbas
- Electrical Engineering Department, College of Engineering, King Khalid University, Abha, 61421, Saudi Arabia
| | - Ben Othman Soufiene
- PRINCE Laboratory Research, ISITcom, Hammam Sousse, University of Sousse, Sousse, Tunisia.
| |
Collapse
|
34
|
Mirasbekov Y, Aidossov N, Mashekova A, Zarikas V, Zhao Y, Ng EYK, Midlenko A. Fully Interpretable Deep Learning Model Using IR Thermal Images for Possible Breast Cancer Cases. Biomimetics (Basel) 2024; 9:609. [PMID: 39451815 PMCID: PMC11506535 DOI: 10.3390/biomimetics9100609] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Revised: 10/04/2024] [Accepted: 10/05/2024] [Indexed: 10/26/2024] Open
Abstract
Breast cancer remains a global health problem requiring effective diagnostic methods for early detection, in order to achieve the World Health Organization's ultimate goal of breast self-examination. A literature review indicates the urgency of improving diagnostic methods and identifies thermography as a promising, cost-effective, non-invasive, adjunctive, and complementary detection method. This research explores the potential of using machine learning techniques, specifically Bayesian networks combined with convolutional neural networks, to improve possible breast cancer diagnosis at early stages. Explainable artificial intelligence aims to clarify the reasoning behind any output of artificial neural network-based models. The proposed integration adds interpretability of the diagnosis, which is particularly significant for a medical diagnosis. We constructed two diagnostic expert models: Model A and Model B. In this research, Model A, combining thermal images after the explainable artificial intelligence process together with medical records, achieved an accuracy of 84.07%, while model B, which also includes a convolutional neural network prediction, achieved an accuracy of 90.93%. These results demonstrate the potential of explainable artificial intelligence to improve possible breast cancer diagnosis, with very high accuracy.
Collapse
Affiliation(s)
- Yerken Mirasbekov
- School of Engineering and Digital Sciences, Nazarbayev University, Astana 010000, Kazakhstan (N.A.); (Y.Z.)
| | - Nurduman Aidossov
- School of Engineering and Digital Sciences, Nazarbayev University, Astana 010000, Kazakhstan (N.A.); (Y.Z.)
| | - Aigerim Mashekova
- School of Engineering and Digital Sciences, Nazarbayev University, Astana 010000, Kazakhstan (N.A.); (Y.Z.)
| | - Vasilios Zarikas
- Department of Mathematics, University of Thessaly, GR-35100 Lamia, Greece;
- Mathematical Sciences Research Laboratory (MSRL), GR-35100 Lamia, Greece
| | - Yong Zhao
- School of Engineering and Digital Sciences, Nazarbayev University, Astana 010000, Kazakhstan (N.A.); (Y.Z.)
| | - Eddie Yin Kwee Ng
- School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore 639798, Singapore;
| | - Anna Midlenko
- School of Medicine, Nazarbayev University, Astana 010000, Kazakhstan;
| |
Collapse
|
35
|
Bandyopadhyay A, Albashayreh A, Zeinali N, Fan W, Gilbertson-White S. Using real-world electronic health record data to predict the development of 12 cancer-related symptoms in the context of multimorbidity. JAMIA Open 2024; 7:ooae082. [PMID: 39282082 PMCID: PMC11397936 DOI: 10.1093/jamiaopen/ooae082] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Revised: 08/09/2024] [Accepted: 09/05/2024] [Indexed: 09/18/2024] Open
Abstract
Objective This study uses electronic health record (EHR) data to predict 12 common cancer symptoms, assessing the efficacy of machine learning (ML) models in identifying symptom influencers. Materials and Methods We analyzed EHR data of 8156 adults diagnosed with cancer who underwent cancer treatment from 2017 to 2020. Structured and unstructured EHR data were sourced from the Enterprise Data Warehouse for Research at the University of Iowa Hospital and Clinics. Several predictive models, including logistic regression, random forest (RF), and XGBoost, were employed to forecast symptom development. The performances of the models were evaluated by F1-score and area under the curve (AUC) on the testing set. The SHapley Additive exPlanations framework was used to interpret these models and identify the predictive risk factors associated with fatigue as an exemplar. Results The RF model exhibited superior performance with a macro average AUC of 0.755 and an F1-score of 0.729 in predicting a range of cancer-related symptoms. For instance, the RF model achieved an AUC of 0.954 and an F1-score of 0.914 for pain prediction. Key predictive factors identified included clinical history, cancer characteristics, treatment modalities, and patient demographics depending on the symptom. For example, the odds ratio (OR) for fatigue was significantly influenced by allergy (OR = 2.3, 95% CI: 1.8-2.9) and colitis (OR = 1.9, 95% CI: 1.5-2.4). Discussion Our research emphasizes the critical integration of multimorbidity and patient characteristics in modeling cancer symptoms, revealing the considerable influence of chronic conditions beyond cancer itself. Conclusion We highlight the potential of ML for predicting cancer symptoms, suggesting a pathway for integrating such models into clinical systems to enhance personalized care and symptom management.
Collapse
Affiliation(s)
- Anindita Bandyopadhyay
- Department of Business Analytics, University of Iowa, Iowa City, IA 52242, United States
| | - Alaa Albashayreh
- College of Nursing, University of Iowa, Iowa City, IA 52242, United States
| | - Nahid Zeinali
- Department of Informatics, University of Iowa, Iowa City, IA 52242, United States
| | - Weiguo Fan
- Department of Business Analytics, University of Iowa, Iowa City, IA 52242, United States
| | | |
Collapse
|
36
|
Abdullakutty F, Akbari Y, Al-Maadeed S, Bouridane A, Talaat IM, Hamoudi R. Histopathology in focus: a review on explainable multi-modal approaches for breast cancer diagnosis. Front Med (Lausanne) 2024; 11:1450103. [PMID: 39403286 PMCID: PMC11471683 DOI: 10.3389/fmed.2024.1450103] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2024] [Accepted: 09/12/2024] [Indexed: 01/11/2025] Open
Abstract
Precision and timeliness in breast cancer detection are paramount for improving patient outcomes. Traditional diagnostic methods have predominantly relied on unimodal approaches, but recent advancements in medical data analytics have enabled the integration of diverse data sources beyond conventional imaging techniques. This review critically examines the transformative potential of integrating histopathology images with genomic data, clinical records, and patient histories to enhance diagnostic accuracy and comprehensiveness in multi-modal diagnostic techniques. It explores early, intermediate, and late fusion methods, as well as advanced deep multimodal fusion techniques, including encoder-decoder architectures, attention-based mechanisms, and graph neural networks. An overview of recent advancements in multimodal tasks such as Visual Question Answering (VQA), report generation, semantic segmentation, and cross-modal retrieval is provided, highlighting the utilization of generative AI and visual language models. Additionally, the review delves into the role of Explainable Artificial Intelligence (XAI) in elucidating the decision-making processes of sophisticated diagnostic algorithms, emphasizing the critical need for transparency and interpretability. By showcasing the importance of explainability, we demonstrate how XAI methods, including Grad-CAM, SHAP, LIME, trainable attention, and image captioning, enhance diagnostic precision, strengthen clinician confidence, and foster patient engagement. The review also discusses the latest XAI developments, such as X-VARs, LeGrad, LangXAI, LVLM-Interpret, and ex-ILP, to demonstrate their potential utility in multimodal breast cancer detection, while identifying key research gaps and proposing future directions for advancing the field.
Collapse
Affiliation(s)
| | - Younes Akbari
- Department of Computer Science and Engineering, Qatar University, Doha, Qatar
| | - Somaya Al-Maadeed
- Department of Computer Science and Engineering, Qatar University, Doha, Qatar
| | - Ahmed Bouridane
- Computer Engineering Department, College of Computing and Informatics, University of Sharjah, Sharjah, United Arab Emirates
| | - Iman M. Talaat
- Clinical Sciences Department, College of Medicine, University of Sharjah, Sharjah, United Arab Emirates
- Research Institute for Medical and Health Sciences, University of Sharjah, Sharjah, United Arab Emirates
| | - Rifat Hamoudi
- Clinical Sciences Department, College of Medicine, University of Sharjah, Sharjah, United Arab Emirates
- Research Institute for Medical and Health Sciences, University of Sharjah, Sharjah, United Arab Emirates
| |
Collapse
|
37
|
Zubair M, Owais M, Mahmood T, Iqbal S, Usman SM, Hussain I. Enhanced gastric cancer classification and quantification interpretable framework using digital histopathology images. Sci Rep 2024; 14:22533. [PMID: 39342030 PMCID: PMC11439054 DOI: 10.1038/s41598-024-73823-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2024] [Accepted: 09/20/2024] [Indexed: 10/01/2024] Open
Abstract
Recent developments have highlighted the critical role that computer-aided diagnosis (CAD) systems play in analyzing whole-slide digital histopathology images for detecting gastric cancer (GC). We present a novel framework for gastric histology classification and segmentation (GHCS) that offers modest yet meaningful improvements over existing CAD models for GC classification and segmentation. Our methodology achieves marginal improvements over conventional deep learning (DL) and machine learning (ML) models by adaptively focusing on pertinent characteristics of images. This contributes significantly to our study, highlighting that the proposed model, which performs well on normalized images, is robust in certain respects, particularly in handling variability and generalizing to different datasets. We anticipate that this robustness will lead to better results across various datasets. An expectation-maximizing Naïve Bayes classifier that uses an updated Gaussian Mixture Model is at the heart of the suggested GHCS framework. The effectiveness of our classifier is demonstrated by experimental validation on two publicly available datasets, which produced exceptional classification accuracies of 98.87% and 97.28% on validation sets and 98.47% and 97.31% on test sets. Our framework shows a slight but consistent improvement over previously existing techniques in gastric histopathology image classification tasks, as demonstrated by comparative analysis. This may be attributed to its ability to capture critical features of gastric histopathology images better. Furthermore, using an improved Fuzzy c-means method, our study produces good results in GC histopathology picture segmentation, outperforming state-of-the-art segmentation models with a Dice coefficient of 65.21% and a Jaccard index of 60.24%. The model's interpretability is complemented by Grad-CAM visualizations, which help understand the decision-making process and increase the model's trustworthiness for end-users, especially clinicians.
Collapse
Affiliation(s)
- Muhammad Zubair
- Faculty of Information Technology & Computer Science, University of Central Punjab, Lahore, Punjab, Pakistan
| | - Muhammad Owais
- Khalifa University Center for Autonomous Robotic Systems (KUCARS) and Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates.
| | - Tahir Mahmood
- Division of Electronics and Electrical Engineering, Dongguk University, Seoul, Korea
| | - Saeed Iqbal
- Faculty of Information Technology & Computer Science, University of Central Punjab, Lahore, Punjab, Pakistan
| | - Syed Muhammad Usman
- Department of Computer Science, School of Engineering and Applied Sciences, Bahria University, Islamabad, Pakistan
| | - Irfan Hussain
- Khalifa University Center for Autonomous Robotic Systems (KUCARS) and Department of Mechanical & Nuclear Engineering, Khalifa University, Abu Dhabi, United Arab Emirates
| |
Collapse
|
38
|
Jiang B, Bao L, He S, Chen X, Jin Z, Ye Y. Deep learning applications in breast cancer histopathological imaging: diagnosis, treatment, and prognosis. Breast Cancer Res 2024; 26:137. [PMID: 39304962 DOI: 10.1186/s13058-024-01895-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 09/16/2024] [Indexed: 09/22/2024] Open
Abstract
Breast cancer is the most common malignant tumor among women worldwide and remains one of the leading causes of death among women. Its incidence and mortality rates are continuously rising. In recent years, with the rapid advancement of deep learning (DL) technology, DL has demonstrated significant potential in breast cancer diagnosis, prognosis evaluation, and treatment response prediction. This paper reviews relevant research progress and applies DL models to image enhancement, segmentation, and classification based on large-scale datasets from TCGA and multiple centers. We employed foundational models such as ResNet50, Transformer, and Hover-net to investigate the performance of DL models in breast cancer diagnosis, treatment, and prognosis prediction. The results indicate that DL techniques have significantly improved diagnostic accuracy and efficiency, particularly in predicting breast cancer metastasis and clinical prognosis. Furthermore, the study emphasizes the crucial role of robust databases in developing highly generalizable models. Future research will focus on addressing challenges related to data management, model interpretability, and regulatory compliance, ultimately aiming to provide more precise clinical treatment and prognostic evaluation programs for breast cancer patients.
Collapse
Affiliation(s)
- Bitao Jiang
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China.
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China.
| | - Lingling Bao
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China
| | - Songqin He
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China
| | - Xiao Chen
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China
| | - Zhihui Jin
- Department of Hematology and Oncology, Beilun District People's Hospital, Ningbo, 315800, China
- Department of Hematology and Oncology, Beilun Branch of the First Affiliated Hospital of Zhejiang University, Ningbo, 315800, China
| | - Yingquan Ye
- Department of Oncology, The 906th Hospital of the Joint Logistics Force of the Chinese People's Liberation Army, Ningbo, 315100, China.
| |
Collapse
|
39
|
Patrick U, Rao SK, Jagan BOL, Rai HM, Agarwal S, Pak W. Radar-Based Target Tracking Using Deep Learning Approaches with Unscented Kalman Filter. APPLIED SCIENCES 2024; 14:8332. [DOI: 10.3390/app14188332] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2025]
Abstract
Machine learning, a rapidly growing field, has attracted numerous researchers for its ability to automatically learn from and make predictions based on data. This manuscript presents an innovative approach to estimating the covariance matrix of noise in radar measurements for target tracking, resulting from collaborative efforts. Traditionally, researchers have assumed that the covariance matrix of noise in sonar measurements is present in the vast majority of literature related to target tracking. On the other hand, this research aims to estimate it by employing deep learning algorithms with noisy measurements in range, bearing, and elevation from radar sensors. This collaborative approach, involving multiple disciplines, provides a more precise and accurate covariance matrix estimate. Additionally, the unscented Kalman filter was combined with the gated recurrent unit, multilayer perceptron, convolutional neural network, and long short-term memory to accomplish the task of 3D target tracking in an airborne environment. The quantification of the results was achieved through the use of Monte Carlo simulations, which demonstrated that the convolutional neural network performed better than any other approach. The system was simulated using a Python program, and the proposed method offers higher accuracy and faster convergence time than conventional target tracking methods. This is a demonstration of the potential that collaboration can have in research.
Collapse
Affiliation(s)
- Uwigize Patrick
- Department of Electronics and Communication Engineering, Koneru Lakshmaiah Education Foundation, Guntur 522302, India
| | - S. Koteswara Rao
- Department of Electronics and Communication Engineering, Koneru Lakshmaiah Education Foundation, Guntur 522302, India
| | - B. Omkar Lakshmi Jagan
- Department of Computer Science and Engineering, Vignan’s Institute of Information Technology, Visakhapatnam 530049, India
| | - Hari Mohan Rai
- Department of Artificial Intelligence and Information Systems, Samarkand State University, University Boulevard 15, Samarkand City 140104, Samarqand Region, Uzbekistan
| | - Saurabh Agarwal
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
| | - Wooguil Pak
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
| |
Collapse
|
40
|
Zhao X, Du Y, Yue H. Skeletal Muscle Segmentation at the Level of the Third Lumbar Vertebra (L3) in Low-Dose Computed Tomography: A Lightweight Algorithm. Tomography 2024; 10:1513-1526. [PMID: 39330757 PMCID: PMC11435900 DOI: 10.3390/tomography10090111] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2024] [Revised: 09/04/2024] [Accepted: 09/09/2024] [Indexed: 09/28/2024] Open
Abstract
BACKGROUND The cross-sectional area of skeletal muscles at the level of the third lumbar vertebra (L3) measured from computed tomography (CT) images is an established imaging biomarker used to assess patients' nutritional status. With the increasing prevalence of low-dose CT scans in clinical practice, accurate and automated skeletal muscle segmentation at the L3 level in low-dose CT images has become an issue to address. This study proposed a lightweight algorithm for automated segmentation of skeletal muscles at the L3 level in low-dose CT images. METHODS This study included 57 patients with rectal cancer, with both low-dose plain and contrast-enhanced pelvic CT image series acquired using a radiotherapy CT scanner. A training set of 30 randomly selected patients was used to develop a lightweight segmentation algorithm, and the other 27 patients were used as the test set. A radiologist selected the most representative axial CT image at the L3 level for both the image series for all the patients, and three groups of observers manually annotated the skeletal muscles in the 54 CT images of the test set as the gold standard. The performance of the proposed algorithm was evaluated in terms of the Dice similarity coefficient (DSC), precision, recall, 95th percentile of the Hausdorff distance (HD95), and average surface distance (ASD). The running time of the proposed algorithm was recorded. An open source deep learning-based AutoMATICA algorithm was compared with the proposed algorithm. The inter-observer variations were also used as the reference. RESULTS The DSC, precision, recall, HD95, ASD, and running time were 93.2 ± 1.9% (mean ± standard deviation), 96.7 ± 2.9%, 90.0 ± 2.9%, 4.8 ± 1.3 mm, 0.8 ± 0.2 mm, and 303 ± 43 ms (on CPU) for the proposed algorithm, and 94.1 ± 4.1%, 92.7 ± 5.5%, 95.7 ± 4.0%, 7.4 ± 5.7 mm, 0.9 ± 0.6 mm, and 448 ± 40 ms (on GPU) for AutoMATICA, respectively. The differences between the proposed algorithm and the inter-observer reference were 4.7%, 1.2%, 7.9%, 3.2 mm, and 0.6 mm, respectively, for the averaged DSC, precision, recall, HD95, and ASD. CONCLUSION The proposed algorithm can be used to segment skeletal muscles at the L3 level in either the plain or enhanced low-dose CT images.
Collapse
Affiliation(s)
- Xuzhi Zhao
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing 100044, China
| | - Yi Du
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing 100142, China
- Institute of Medical Technology, Peking University Health Science Center, Beijing 100191, China
| | - Haizhen Yue
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing 100142, China
| |
Collapse
|
41
|
Rai HM, Dashkevych S, Yoo J. Next-Generation Diagnostics: The Impact of Synthetic Data Generation on the Detection of Breast Cancer from Ultrasound Imaging. MATHEMATICS 2024; 12:2808. [DOI: 10.3390/math12182808] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2025]
Abstract
Breast cancer is one of the most lethal and widespread diseases affecting women worldwide. As a result, it is necessary to diagnose breast cancer accurately and efficiently utilizing the most cost-effective and widely used methods. In this research, we demonstrated that synthetically created high-quality ultrasound data outperformed conventional augmentation strategies for efficiently diagnosing breast cancer using deep learning. We trained a deep-learning model using the EfficientNet-B7 architecture and a large dataset of 3186 ultrasound images acquired from multiple publicly available sources, as well as 10,000 synthetically generated images using generative adversarial networks (StyleGAN3). The model was trained using five-fold cross-validation techniques and validated using four metrics: accuracy, recall, precision, and the F1 score measure. The results showed that integrating synthetically produced data into the training set increased the classification accuracy from 88.72% to 92.01% based on the F1 score, demonstrating the power of generative models to expand and improve the quality of training datasets in medical-imaging applications. This demonstrated that training the model using a larger set of data comprising synthetic images significantly improved its performance by more than 3% over the genuine dataset with common augmentation. Various data augmentation procedures were also investigated to improve the training set’s diversity and representativeness. This research emphasizes the relevance of using modern artificial intelligence and machine-learning technologies in medical imaging by providing an effective strategy for categorizing ultrasound images, which may lead to increased diagnostic accuracy and optimal treatment options. The proposed techniques are highly promising and have strong potential for future clinical application in the diagnosis of breast cancer.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Gyeonggi-do, Republic of Korea
| | - Serhii Dashkevych
- Department of Computer Engineering, Vistula University, Stokłosy 3, 02-787 Warszawa, Poland
| | - Joon Yoo
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Gyeonggi-do, Republic of Korea
| |
Collapse
|
42
|
Ullah Z, Jamjoom M, Thirumalaisamy M, Alajmani SH, Saleem F, Sheikh-Akbari A, Khan UA. A Deep Learning Based Intelligent Decision Support System for Automatic Detection of Brain Tumor. Biomed Eng Comput Biol 2024; 15:11795972241277322. [PMID: 39238891 PMCID: PMC11375672 DOI: 10.1177/11795972241277322] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Accepted: 08/06/2024] [Indexed: 09/07/2024] Open
Abstract
Brain tumor (BT) is an awful disease and one of the foremost causes of death in human beings. BT develops mainly in 2 stages and varies by volume, form, and structure, and can be cured with special clinical procedures such as chemotherapy, radiotherapy, and surgical mediation. With revolutionary advancements in radiomics and research in medical imaging in the past few years, computer-aided diagnostic systems (CAD), especially deep learning, have played a key role in the automatic detection and diagnosing of various diseases and significantly provided accurate decision support systems for medical clinicians. Thus, convolution neural network (CNN) is a commonly utilized methodology developed for detecting various diseases from medical images because it is capable of extracting distinct features from an image under investigation. In this study, a deep learning approach is utilized to extricate distinct features from brain images in order to detect BT. Hence, CNN from scratch and transfer learning models (VGG-16, VGG-19, and LeNet-5) are developed and tested on brain images to build an intelligent decision support system for detecting BT. Since deep learning models require large volumes of data, data augmentation is used to populate the existing dataset synthetically in order to utilize the best fit detecting models. Hyperparameter tuning was conducted to set the optimum parameters for training the models. The achieved results show that VGG models outperformed others with an accuracy rate of 99.24%, average precision of 99%, average recall of 99%, average specificity of 99%, and average f1-score of 99% each. The results of the proposed models compared to the other state-of-the-art models in the literature show better performance of the proposed models in terms of accuracy, sensitivity, specificity, and f1-score. Moreover, comparative analysis shows that the proposed models are reliable in that they can be used for detecting BT as well as helping medical practitioners to diagnose BT.
Collapse
Affiliation(s)
- Zahid Ullah
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University, Riyadh, Saudi Arabia
| | - Mona Jamjoom
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
| | | | - Samah H Alajmani
- Department of Information Technology, College of Computers and Information Technology, Taif University, Taif, Saudi Arabia
| | - Farrukh Saleem
- School of Built Environment, Engineering, and Computing, Leeds Beckett University, Leeds, UK
| | - Akbar Sheikh-Akbari
- School of Built Environment, Engineering, and Computing, Leeds Beckett University, Leeds, UK
| | - Usman Ali Khan
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia
| |
Collapse
|
43
|
Jiang M, Bian F, Zhang J, Pu Z, Li H, Zhang Y, Chu Y, Fan Y, Jiang J. An Automatic Coronary Microvascular Dysfunction Classification Method Based on Hybrid ECG Features and Expert Features. IEEE J Biomed Health Inform 2024; 28:5103-5112. [PMID: 38923474 DOI: 10.1109/jbhi.2024.3419090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2024]
Abstract
OBJECTIVE In recent years, the early diagnosis and treatment of coronary microvascular dysfunction (CMD) have become crucial for preventing coronary heart disease. This paper aims to develop a computer-assisted autonomous diagnosis method for CMD by using ECG features and expert features. APPROACH Clinical electrocardiogram (ECG), myocardial contrast echocardiography (MCE), and coronary angiography (CAG) are used in our method. Firstly, morphological features, temporal features, and T-wave features of ECG are extracted by multi-channel residual network with BiLSTM (MCResnet-BiLSTM) model and the multi-source T-wave features (MTF) extraction model, respectively. And these features are fused to form ECG features. In addition, the CFR[Formula: see text] is calculated based on the parameters related to the MCE at rest and stress state, and the Angio-IMR is calculated based on CAG. The combination of CFR[Formula: see text] and Angio-IMR is termed as expert features. Furthermore, the hybrid features, fused from the ECG features and the expert features, are input into the multilayer perceptron to implement the identification of CMD. And the weighted sum of the softmax loss and center loss is used as the total loss function for training the classification model, which optimizes the classification ability of the model. RESULT The proposed method achieved 93.36% accuracy, 94.46% specificity, 92.10% sensitivity, 95.89% precision, and 93.95% F1 score on the clinical dataset of the Second Affiliated Hospital of Zhejiang University. CONCLUSION The proposed method accurately extracts global ECG features, combines them with expert features to obtain hybrid features, and uses weighted loss to significantly improve diagnostic accuracy. It provides a novel and practical method for the clinical diagnosis of CMD.
Collapse
|
44
|
Iqbal MS, Belal Bin Heyat M, Parveen S, Ammar Bin Hayat M, Roshanzamir M, Alizadehsani R, Akhtar F, Sayeed E, Hussain S, Hussein HS, Sawan M. Progress and trends in neurological disorders research based on deep learning. Comput Med Imaging Graph 2024; 116:102400. [PMID: 38851079 DOI: 10.1016/j.compmedimag.2024.102400] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Revised: 05/07/2024] [Accepted: 05/13/2024] [Indexed: 06/10/2024]
Abstract
In recent years, deep learning (DL) has emerged as a powerful tool in clinical imaging, offering unprecedented opportunities for the diagnosis and treatment of neurological disorders (NDs). This comprehensive review explores the multifaceted role of DL techniques in leveraging vast datasets to advance our understanding of NDs and improve clinical outcomes. Beginning with a systematic literature review, we delve into the utilization of DL, particularly focusing on multimodal neuroimaging data analysis-a domain that has witnessed rapid progress and garnered significant scientific interest. Our study categorizes and critically analyses numerous DL models, including Convolutional Neural Networks (CNNs), LSTM-CNN, GAN, and VGG, to understand their performance across different types of Neurology Diseases. Through particular analysis, we identify key benchmarks and datasets utilized in training and testing DL models, shedding light on the challenges and opportunities in clinical neuroimaging research. Moreover, we discuss the effectiveness of DL in real-world clinical scenarios, emphasizing its potential to revolutionize ND diagnosis and therapy. By synthesizing existing literature and describing future directions, this review not only provides insights into the current state of DL applications in ND analysis but also covers the way for the development of more efficient and accessible DL techniques. Finally, our findings underscore the transformative impact of DL in reshaping the landscape of clinical neuroimaging, offering hope for enhanced patient care and groundbreaking discoveries in the field of neurology. This review paper is beneficial for neuropathologists and new researchers in this field.
Collapse
Affiliation(s)
- Muhammad Shahid Iqbal
- Department of Computer Science and Information Technology, Women University of Azad Jammu & Kashmir, Bagh, Pakistan.
| | - Md Belal Bin Heyat
- CenBRAIN Neurotech Center of Excellence, School of Engineering, Westlake University, Hangzhou, Zhejiang, China.
| | - Saba Parveen
- College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China.
| | | | - Mohamad Roshanzamir
- Department of Computer Engineering, Faculty of Engineering, Fasa University, Fasa, Iran.
| | - Roohallah Alizadehsani
- Institute for Intelligent Systems Research and Innovation, Deakin University, VIC 3216, Australia.
| | - Faijan Akhtar
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China.
| | - Eram Sayeed
- Kisan Inter College, Dhaurahara, Kushinagar, India.
| | - Sadiq Hussain
- Department of Examination, Dibrugarh University, Assam 786004, India.
| | - Hany S Hussein
- Electrical Engineering Department, Faculty of Engineering, King Khalid University, Abha 61411, Saudi Arabia; Electrical Engineering Department, Faculty of Engineering, Aswan University, Aswan 81528, Egypt.
| | - Mohamad Sawan
- CenBRAIN Neurotech Center of Excellence, School of Engineering, Westlake University, Hangzhou, Zhejiang, China.
| |
Collapse
|
45
|
Rai HM, Yoo J, Dashkevych S. GAN-SkipNet: A Solution for Data Imbalance in Cardiac Arrhythmia Detection Using Electrocardiogram Signals from a Benchmark Dataset. MATHEMATICS 2024; 12:2693. [DOI: 10.3390/math12172693] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2025]
Abstract
Electrocardiography (ECG) plays a pivotal role in monitoring cardiac health, yet the manual analysis of ECG signals is challenging due to the complex task of identifying and categorizing various waveforms and morphologies within the data. Additionally, ECG datasets often suffer from a significant class imbalance issue, which can lead to inaccuracies in detecting minority class samples. To address these challenges and enhance the effectiveness and efficiency of cardiac arrhythmia detection from imbalanced ECG datasets, this study proposes a novel approach. This research leverages the MIT-BIH arrhythmia dataset, encompassing a total of 109,446 ECG beats distributed across five classes following the Association for the Advancement of Medical Instrumentation (AAMI) standard. Given the dataset’s inherent class imbalance, a 1D generative adversarial network (GAN) model is introduced, incorporating the Bi-LSTM model to synthetically generate the two minority signal classes, which represent a mere 0.73% fusion (F) and 2.54% supraventricular (S) of the data. The generated signals are rigorously evaluated for similarity to real ECG data using three key metrics: mean squared error (MSE), structural similarity index (SSIM), and Pearson correlation coefficient (r). In addition to addressing data imbalance, the work presents three deep learning models tailored for ECG classification: SkipCNN (a convolutional neural network with skip connections), SkipCNN+LSTM, and SkipCNN+LSTM+Attention mechanisms. To further enhance efficiency and accuracy, the test dataset is rigorously assessed using an ensemble model, which consistently outperforms the individual models. The performance evaluation employs standard metrics such as precision, recall, and F1-score, along with their average, macro average, and weighted average counterparts. Notably, the SkipCNN+LSTM model emerges as the most promising, achieving remarkable precision, recall, and F1-scores of 99.3%, which were further elevated to an impressive 99.60% through ensemble techniques. Consequently, with this innovative combination of data balancing techniques, the GAN-SkipNet model not only resolves the challenges posed by imbalanced data but also provides a robust and reliable solution for cardiac arrhythmia detection. This model stands poised for clinical applications, offering the potential to be deployed in hospitals for real-time cardiac arrhythmia detection, thereby benefiting patients and healthcare practitioners alike.
Collapse
Affiliation(s)
- Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
| | - Joon Yoo
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
| | - Serhii Dashkevych
- Department of Computer Science, Data Scientist, Vistula University, Stokłosy 3, 02-787 Warszawa, Poland
| |
Collapse
|
46
|
Moqurrab SA, Rai HM, Yoo J. HRIDM: Hybrid Residual/Inception-Based Deeper Model for Arrhythmia Detection from Large Sets of 12-Lead ECG Recordings. ALGORITHMS 2024; 17:364. [DOI: 10.3390/a17080364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/11/2024]
Abstract
Heart diseases such as cardiovascular and myocardial infarction are the foremost reasons of death in the world. The timely, accurate, and effective prediction of heart diseases is crucial for saving lives. Electrocardiography (ECG) is a primary non-invasive method to identify cardiac abnormalities. However, manual interpretation of ECG recordings for heart disease diagnosis is a time-consuming and inaccurate process. For the accurate and efficient detection of heart diseases from the 12-lead ECG dataset, we have proposed a hybrid residual/inception-based deeper model (HRIDM). In this study, we have utilized ECG datasets from various sources, which are multi-institutional large ECG datasets. The proposed model is trained on 12-lead ECG data from over 10,000 patients. We have compared the proposed model with several state-of-the-art (SOTA) models, such as LeNet-5, AlexNet, VGG-16, ResNet-50, Inception, and LSTM, on the same training and test datasets. To show the effectiveness of the computational efficiency of the proposed model, we have only trained over 20 epochs without GPU support and we achieved an accuracy of 50.87% on the test dataset for 27 categories of heart abnormalities. We found that our proposed model outperformed the previous studies which participated in the official PhysioNet/CinC Challenge 2020 and achieved fourth place as compared with the 41 official ranking teams. The result of this study indicates that the proposed model is an implying new method for predicting heart diseases using 12-lead ECGs.
Collapse
Affiliation(s)
- Syed Atif Moqurrab
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
| | - Hari Mohan Rai
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
| | - Joon Yoo
- School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
| |
Collapse
|
47
|
Mudavadkar GR, Deng M, Al-Heejawi SMA, Arora IH, Breggia A, Ahmad B, Christman R, Ryan ST, Amal S. Gastric Cancer Detection with Ensemble Learning on Digital Pathology: Use Case of Gastric Cancer on GasHisSDB Dataset. Diagnostics (Basel) 2024; 14:1746. [PMID: 39202233 PMCID: PMC11354078 DOI: 10.3390/diagnostics14161746] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2024] [Revised: 08/01/2024] [Accepted: 08/02/2024] [Indexed: 09/03/2024] Open
Abstract
Gastric cancer has become a serious worldwide health concern, emphasizing the crucial importance of early diagnosis measures to improve patient outcomes. While traditional histological image analysis is regarded as the clinical gold standard, it is labour intensive and manual. In recognition of this problem, there has been a rise in interest in the use of computer-aided diagnostic tools to help pathologists with their diagnostic efforts. In particular, deep learning (DL) has emerged as a promising solution in this sector. However, current DL models are still restricted in their ability to extract extensive visual characteristics for correct categorization. To address this limitation, this study proposes the use of ensemble models, which incorporate the capabilities of several deep-learning architectures and use aggregate knowledge of many models to improve classification performance, allowing for more accurate and efficient gastric cancer detection. To determine how well these proposed models performed, this study compared them with other works, all of which were based on the Gastric Histopathology Sub-Size Images Database, a publicly available dataset for gastric cancer. This research demonstrates that the ensemble models achieved a high detection accuracy across all sub-databases, with an average accuracy exceeding 99%. Specifically, ResNet50, VGGNet, and ResNet34 performed better than EfficientNet and VitNet. For the 80 × 80-pixel sub-database, ResNet34 exhibited an accuracy of approximately 93%, VGGNet achieved 94%, and the ensemble model excelled with 99%. In the 120 × 120-pixel sub-database, the ensemble model showed 99% accuracy, VGGNet 97%, and ResNet50 approximately 97%. For the 160 × 160-pixel sub-database, the ensemble model again achieved 99% accuracy, VGGNet 98%, ResNet50 98%, and EfficientNet 92%, highlighting the ensemble model's superior performance across all resolutions. Overall, the ensemble model consistently provided an accuracy of 99% across the three sub-pixel categories. These findings show that ensemble models may successfully detect critical characteristics from smaller patches and achieve high performance. The findings will help pathologists diagnose gastric cancer using histopathological images, leading to earlier identification and higher patient survival rates.
Collapse
Affiliation(s)
- Govind Rajesh Mudavadkar
- College of Engineering, Northeastern University, Boston, MA 02115, USA; (G.R.M.); (M.D.); (S.M.A.A.-H.)
| | - Mo Deng
- College of Engineering, Northeastern University, Boston, MA 02115, USA; (G.R.M.); (M.D.); (S.M.A.A.-H.)
| | | | - Isha Hemant Arora
- Khoury College of Computer Sciences, Northeastern University, Boston, MA 02115, USA;
| | - Anne Breggia
- MaineHealth Institute for Research, Scarborough, ME 04074, USA
| | - Bilal Ahmad
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Robert Christman
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Stephen T. Ryan
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Saeed Amal
- The Roux Institute, Department of Bioengineering, College of Engineering at Northeastern University, Boston, MA 02115, USA
| |
Collapse
|
48
|
Ma L, Zhang F. A Novel Real-Time Detection and Classification Method for ECG Signal Images Based on Deep Learning. SENSORS (BASEL, SWITZERLAND) 2024; 24:5087. [PMID: 39204785 PMCID: PMC11360666 DOI: 10.3390/s24165087] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/15/2024] [Revised: 07/28/2024] [Accepted: 08/02/2024] [Indexed: 09/04/2024]
Abstract
In this paper, a novel deep learning method Mamba-RAYOLO is presented, which can improve detection and classification in the processing and analysis of ECG images in real time by integrating three advanced modules. The feature extraction module in our work with a multi-branch structure during training can capture a wide range of features to ensure efficient inference and rich feature extraction. The attention mechanism module utilized in our proposed network can dynamically focus on the most relevant spatial and channel-wise features to improve detection accuracy and computational efficiency. Then, the extracted features can be refined for efficient spatial feature processing and robust feature fusion. Several sets of experiments have been carried out to test the validity of the proposed Mamba-RAYOLO and these indicate that our method has made significant improvements in the detection and classification of ECG images. The research offers a promising framework for more accurate and efficient medical ECG diagnostics.
Collapse
Affiliation(s)
- Linjuan Ma
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China;
| | - Fuquan Zhang
- College of Computer and Control Engineering, Minjiang University, Fuzhou 350108, China
| |
Collapse
|
49
|
Ding X, Jiang X, Zheng H, Shi H, Wang B, Chan S. MARes-Net: multi-scale attention residual network for jaw cyst image segmentation. Front Bioeng Biotechnol 2024; 12:1454728. [PMID: 39161348 PMCID: PMC11330813 DOI: 10.3389/fbioe.2024.1454728] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Accepted: 07/25/2024] [Indexed: 08/21/2024] Open
Abstract
Jaw cyst is a fluid-containing cystic lesion that can occur in any part of the jaw and cause facial swelling, dental lesions, jaw fractures, and other associated issues. Due to the diversity and complexity of jaw images, existing deep-learning methods still have challenges in segmentation. To this end, we propose MARes-Net, an innovative multi-scale attentional residual network architecture. Firstly, the residual connection is used to optimize the encoder-decoder process, which effectively solves the gradient disappearance problem and improves the training efficiency and optimization ability. Secondly, the scale-aware feature extraction module (SFEM) significantly enhances the network's perceptual abilities by extending its receptive field across various scales, spaces, and channel dimensions. Thirdly, the multi-scale compression excitation module (MCEM) compresses and excites the feature map, and combines it with contextual information to obtain better model performance capabilities. Furthermore, the introduction of the attention gate module marks a significant advancement in refining the feature map output. Finally, rigorous experimentation conducted on the original jaw cyst dataset provided by Quzhou People's Hospital to verify the validity of MARes-Net architecture. The experimental data showed that precision, recall, IoU and F1-score of MARes-Net reached 93.84%, 93.70%, 86.17%, and 93.21%, respectively. Compared with existing models, our MARes-Net shows its unparalleled capabilities in accurately delineating and localizing anatomical structures in the jaw cyst image segmentation.
Collapse
Affiliation(s)
- Xiaokang Ding
- College of Mechanical Engineering, Quzhou University, Quzhou, China
| | - Xiaoliang Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou, China
| | - Huixia Zheng
- Department of Stomatology, Quzhou People’s Hospital, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou, China
| | - Hualuo Shi
- College of Mechanical Engineering, Quzhou University, Quzhou, China
- School of Mechanical Engineering, Hangzhou Dianzi University, Hangzhou, China
| | - Ban Wang
- School of Mechanical Engineering, Hangzhou Dianzi University, Hangzhou, China
| | - Sixian Chan
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China
| |
Collapse
|
50
|
Muthukrishnan V, Jaipurkar S, Damodaran N. Continuum topological derivative - a novel application tool for denoising CT and MRI medical images. BMC Med Imaging 2024; 24:182. [PMID: 39048968 PMCID: PMC11267933 DOI: 10.1186/s12880-024-01341-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 06/18/2024] [Indexed: 07/27/2024] Open
Abstract
BACKGROUND CT and MRI modalities are important diagnostics tools for exploring the anatomical and tissue properties, respectively of the human beings. Several advancements like HRCT, FLAIR and Propeller have advantages in diagnosing the diseases very accurately, but still have enough space for improvements due to the presence of inherent and instrument noises. In the case of CT and MRI, the quantum mottle and the Gaussian and Rayleigh noises, respectively are still present in their advanced modalities of imaging. This paper addresses the denoising problem with continuum topological derivative technique and proved its trustworthiness based on the comparative study with other traditional filtration methods such as spatial, adaptive, frequency and transformation techniques using measures like visual inspection and performance metrics. METHODS This research study focuses on identifying a novel method for denoising by testing different filters on HRCT (High-Resolution Computed Tomography) and MR (Magnetic Resonance) images. The images were acquired from the Image Art Radiological Scan Centre using the SOMATOM CT and SIGNA Explorer (operating at 1.5 Tesla) machines. To compare the performance of the proposed CTD (Continuum Topological Derivative) method, various filters were tested on both HRCT and MR images. The filters tested for comparison were Gaussian (2D convolution operator), Wiener (deconvolution operator), Laplacian and Laplacian diagonal (2nd order partial differential operator), Average, Minimum, and Median (ordinary spatial operators), PMAD (Anisotropic diffusion operator), Kuan (statistical operator), Frost (exponential convolution operator), and HAAR Wavelet (time-frequency operator). The purpose of the study was to evaluate the effectiveness of the CTD method in removing noise compared to the other filters. The performance metrics were analyzed to assess the diligence of noise removal achieved by the CTD method. The primary outcome of the study was the removal of quantum mottle noise in HRCT images, while the secondary outcome focused on removing Gaussian (foreground) and Rayleigh (background) noise in MR images. The study aimed to observe the dynamics of noise removal by examining the values of the performance metrics. In summary, this study aimed to assess the denoising ability of various filters in HRCT and MR images, with the CTD method being the proposed approach. The study evaluated the performance of each filter using specific metrics and compared the results to determine the effectiveness of the CTD method in removing noise from the images. RESULTS Based on the calculated performance metric values, it has been observed that the CTD method successfully removed quantum mottle noise in HRCT images and Gaussian as well as Rayleigh noise in MRI. This can be evidenced by the PSNR (Peak Signal-to-Noise Ratio) metric, which consistently exhibited values ranging from 50 to 65 for all the tested images. Additionally, the CTD method demonstrated remarkably low residual values, typically on the order of e-09, which is a distinctive characteristic across all the images. Furthermore, the performance metrics of the CTD method consistently outperformed those of the other tested methods. Consequently, the results of this study have significant implications for the quality, structural similarity, and contrast of HRCT and MR images, enabling clinicians to obtain finer details for diagnostic purposes. CONCLUSION Continuum topological derivative algorithm is found to be constructive in removing prominent noises in both CT and MRI images and can serve as a potential tool for recognition of anatomical details in case of diseased and normal ones. The results obtained from this research work are highly inspiring and offer great promise in obtaining accurate diagnostic information for critical cases such as Thoracic Cavity Carina, Brain SPI Globe Lens 4th Ventricle, Brain-Middle Cerebral Artery, Brain-Middle Cerebral Artery and neoplastic lesions. These findings lay the foundation for implementing the proposed CTD technique in routine clinical diagnosis.
Collapse
Affiliation(s)
- Viswanath Muthukrishnan
- Central Instrumentation & Service Laboratory, Guindy Campus, University of Madras, Chennai, India
| | | | - Nedumaran Damodaran
- Central Instrumentation & Service Laboratory, Guindy Campus, University of Madras, Chennai, India.
| |
Collapse
|