1
|
Anber B, Yurtkan K. Fractional differentiation based image enhancement for automatic detection of malignant melanoma. BMC Med Imaging 2024; 24:231. [PMID: 39223468 DOI: 10.1186/s12880-024-01400-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2024] [Accepted: 08/14/2024] [Indexed: 09/04/2024] Open
Abstract
Recent improvements in artificial intelligence and computer vision make it possible to automatically detect abnormalities in medical images. Skin lesions are one broad class of them. There are types of lesions that cause skin cancer, again with several types. Melanoma is one of the deadliest types of skin cancer. Its early diagnosis is at utmost importance. The treatments are greatly aided with artificial intelligence by the quick and precise diagnosis of these conditions. The identification and delineation of boundaries inside skin lesions have shown promise when using the basic image processing approaches for edge detection. Further enhancements regarding edge detections are possible. In this paper, the use of fractional differentiation for improved edge detection is explored on the application of skin lesion detection. A framework based on fractional differential filters for edge detection in skin lesion images is proposed that can improve automatic detection rate of malignant melanoma. The derived images are used to enhance the input images. Obtained images then undergo a classification process based on deep learning. A well-studied dataset of HAM10000 is used in the experiments. The system achieves 81.04% accuracy with EfficientNet model using the proposed fractional derivative based enhancements whereas accuracies are around 77.94% when using original images. In almost all the experiments, the enhanced images improved the accuracy. The results show that the proposed method improves the recognition performance.
Collapse
Affiliation(s)
- Basmah Anber
- Computer Engineering Department, Faculty of Engineering, Cyprus International University, via Mersin10, Nicosia, Northern Cyprus, Turkey.
| | - Kamil Yurtkan
- Computer Engineering Department, Faculty of Engineering, Cyprus International University, via Mersin10, Nicosia, Northern Cyprus, Turkey
- Artificial Intelligence Application and Research Center, Cyprus International University, via Mersin10, Nicosia, Northern Cyprus, Turkey
| |
Collapse
|
2
|
Wei D, Jiang Y, Zhou X, Wu D, Feng X. A Review of Advancements and Challenges in Liver Segmentation. J Imaging 2024; 10:202. [PMID: 39194991 DOI: 10.3390/jimaging10080202] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2024] [Revised: 08/07/2024] [Accepted: 08/13/2024] [Indexed: 08/29/2024] Open
Abstract
Liver segmentation technologies play vital roles in clinical diagnosis, disease monitoring, and surgical planning due to the complex anatomical structure and physiological functions of the liver. This paper provides a comprehensive review of the developments, challenges, and future directions in liver segmentation technology. We systematically analyzed high-quality research published between 2014 and 2024, focusing on liver segmentation methods, public datasets, and evaluation metrics. This review highlights the transition from manual to semi-automatic and fully automatic segmentation methods, describes the capabilities and limitations of available technologies, and provides future outlooks.
Collapse
Affiliation(s)
- Di Wei
- Department of Radiology, The Eighth Affiliated Hospital of The Sun Yat-sen University, No. 3025, Middle Shennan Road, Shenzhen 518033, China
| | - Yundan Jiang
- Department of Radiology, The Eighth Affiliated Hospital of The Sun Yat-sen University, No. 3025, Middle Shennan Road, Shenzhen 518033, China
| | - Xuhui Zhou
- Department of Radiology, The Eighth Affiliated Hospital of The Sun Yat-sen University, No. 3025, Middle Shennan Road, Shenzhen 518033, China
| | - Di Wu
- Department of Radiology, The Eighth Affiliated Hospital of The Sun Yat-sen University, No. 3025, Middle Shennan Road, Shenzhen 518033, China
| | - Xiaorong Feng
- Department of Radiology, The Eighth Affiliated Hospital of The Sun Yat-sen University, No. 3025, Middle Shennan Road, Shenzhen 518033, China
| |
Collapse
|
3
|
Fan N, Chen X, Li Y, Zhu Z, Chen X, Yang Z, Yang J. Dual-energy computed tomography with new virtual monoenergetic image reconstruction enhances prostate lesion image quality and improves the diagnostic efficacy for prostate cancer. BMC Med Imaging 2024; 24:212. [PMID: 39134937 PMCID: PMC11321013 DOI: 10.1186/s12880-024-01393-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2024] [Accepted: 08/05/2024] [Indexed: 08/15/2024] Open
Abstract
BACKGROUND Prostate cancer is one of the most common malignant tumors in middle-aged and elderly men and carries significant prognostic implications, and recent studies suggest that dual-energy computed tomography (DECT) utilizing new virtual monoenergetic images can enhance cancer detection rates. This study aimed to assess the impact of virtual monoenergetic images reconstructed from DECT arterial phase scans on the image quality of prostate lesions and their diagnostic performance for prostate cancer. METHODS We conducted a retrospective analysis of 83 patients with prostate cancer or prostatic hyperplasia who underwent DECT scans at Meizhou People's Hospital between July 2019 and December 2023. The variables analyzed included age, tumor diameter and serum prostate-specific antigen (PSA) levels, among others. We also compared CT values, signal-to-noise ratio (SNR), subjective image quality ratings, and contrast-to-noise ratio (CNR) between virtual monoenergetic images (40-100 keV) and conventional linear blending images. Receiver operating characteristic (ROC) curve analyses were performed to evaluate the diagnostic efficacy of virtual monoenergetic images (40 keV and 50 keV) compared to conventional images. RESULTS Virtual monoenergetic images at 40 keV showed significantly higher CT values (168.19 ± 57.14) compared to conventional linear blending images (66.66 ± 15.5) for prostate cancer (P < 0.001). The 50 keV images also demonstrated elevated CT values (121.73 ± 39.21) compared to conventional images (P < 0.001). CNR values for the 40 keV (3.81 ± 2.13) and 50 keV (2.95 ± 1.50) groups were significantly higher than the conventional blending group (P < 0.001). Subjective evaluations indicated markedly better image quality scores for 40 keV (median score of 5) and 50 keV (median score of 5) images compared to conventional images (P < 0.05). ROC curve analysis revealed superior diagnostic accuracy for 40 keV (AUC: 0.910) and 50 keV (AUC: 0.910) images based on CT values compared to conventional images (AUC: 0.849). CONCLUSIONS Virtual monoenergetic images reconstructed at 40 keV and 50 keV from DECT arterial phase scans substantially enhance the image quality of prostate lesions and improve diagnostic efficacy for prostate cancer.
Collapse
Affiliation(s)
- Nina Fan
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China
| | - Xiaofeng Chen
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China
| | - Yulin Li
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China
| | - Zhiqiang Zhu
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China
| | - Xiangguang Chen
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China
| | - Zhiqi Yang
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China.
| | - Jiada Yang
- Department of Radiology, Meizhou People's Hospital, Meizhou, 514000, Guangdong, China.
| |
Collapse
|
4
|
Radhabai PR, Kvn K, Shanmugam A, Imoize AL. An effective no-reference image quality index prediction with a hybrid Artificial Intelligence approach for denoised MRI images. BMC Med Imaging 2024; 24:208. [PMID: 39134983 PMCID: PMC11318287 DOI: 10.1186/s12880-024-01387-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2024] [Accepted: 08/01/2024] [Indexed: 08/16/2024] Open
Abstract
As the quantity and significance of digital pictures in the medical industry continue to increase, Image Quality Assessment (IQA) has recently become a prevalent subject in the research community. Due to the wide range of distortions that Magnetic Resonance Images (MRI) can experience and the wide variety of information they contain, No-Reference Image Quality Assessment (NR-IQA) has always been a challenging study issue. In an attempt to address this issue, a novel hybrid Artificial Intelligence (AI) is proposed to analyze NR-IQ in massive MRI data. First, the features from the denoised MRI images are extracted using the gray level run length matrix (GLRLM) and EfficientNet B7 algorithm. Next, the Multi-Objective Reptile Search Algorithm (MRSA) was proposed for optimal feature vector selection. Then, the Self-evolving Deep Belief Fuzzy Neural network (SDBFN) algorithm was proposed for the effective NR-IQ analysis. The implementation of this research is executed using MATLAB software. The simulation results are compared with the various conventional methods in terms of correlation coefficient (PLCC), Root Mean Square Error (RMSE), Spearman Rank Order Correlation Coefficient (SROCC) and Kendall Rank Order Correlation Coefficient (KROCC), and Mean Absolute Error (MAE). In addition, our proposed approach yielded a quality number approximately we achieved significant 20% improvement than existing methods, with the PLCC parameter showing a notable increase compared to current techniques. Moreover, the RMSE number decreased by 12% when compared to existing methods. Graphical representations indicated mean MAE values of 0.02 for MRI knee dataset, 0.09 for MRI brain dataset, and 0.098 for MRI breast dataset, showcasing significantly lower MAE values compared to the baseline models.
Collapse
Affiliation(s)
| | - Kavitha Kvn
- Department of Communication Engineering, School of Electronics Engineering, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| | - Ashok Shanmugam
- Department of Electronics and Communication Engineering, Vel Tech Multi Tech Dr. Rangarajan Dr. Sakunthala Engineering College, Chennai, Tamil Nadu, India
| | - Agbotiname Lucky Imoize
- Department of Electrical and Electronics Engineering, Faculty of Engineering, University of Lagos, Akoka, Lagos, 100213, Nigeria.
| |
Collapse
|
5
|
Mou E, Wang H, Chen X, Li Z, Cao E, Chen Y, Huang Z, Pang Y. Retinex theory-based nonlinear luminance enhancement and denoising for low-light endoscopic images. BMC Med Imaging 2024; 24:207. [PMID: 39123136 PMCID: PMC11316405 DOI: 10.1186/s12880-024-01386-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2024] [Accepted: 08/01/2024] [Indexed: 08/12/2024] Open
Abstract
BACKGROUND The quality of low-light endoscopic images involves applications in medical disciplines such as physiology and anatomy for the identification and judgement of tissue structures. Due to the use of point light sources and the constraints of narrow physiological structures, medical endoscopic images display uneven brightness, low contrast, and a lack of texture information, presenting diagnostic challenges for physicians. METHODS In this paper, a nonlinear brightness enhancement and denoising network based on Retinex theory is designed to improve the brightness and details of low-light endoscopic images. The nonlinear luminance enhancement module uses higher-order curvilinear functions to improve overall brightness; the dual-attention denoising module captures detailed features of anatomical structures; and the color loss function mitigates color distortion. RESULTS Experimental results on the Endo4IE dataset demonstrate that the proposed method outperforms existing state-of-the-art methods in terms of Peak Signal-to-Noise Ratio (PSNR), Structural Similarity (SSIM), and Learned Perceptual Image Patch Similarity (LPIPS). The PSNR is 27.2202, SSIM is 0.8342, and the LPIPS is 0.1492. It provides a method to enhance image quality in clinical diagnosis and treatment. CONCLUSIONS It offers an efficient method to enhance images captured by endoscopes and offers valuable insights into intricate human physiological structures, which can effectively assist clinical diagnosis and treatment.
Collapse
Affiliation(s)
- En Mou
- School of Communications and Information Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
- School of Medical Information and Engineering, Southwest Medical University, Luzhou, 646000, China
| | - Huiqian Wang
- School of Optoelectronic Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China.
| | - Xiaodong Chen
- School of Medical Information and Engineering, Southwest Medical University, Luzhou, 646000, China
| | - Zhangyong Li
- School of Optoelectronic Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
| | - Enling Cao
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
| | - Yuanyuan Chen
- School of Communications and Information Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
- College of Physics and Telecommunication Engineering, Zhoukou Normal University, Zhoukou, 466001, China
| | - Zhiwei Huang
- School of Medical Information and Engineering, Southwest Medical University, Luzhou, 646000, China
| | - Yu Pang
- School of Optoelectronic Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China.
| |
Collapse
|
6
|
Xie K, Gao L, Zhang Y, Zhang H, Sun J, Lin T, Sui J, Ni X. Metal implant segmentation in CT images based on diffusion model. BMC Med Imaging 2024; 24:204. [PMID: 39107679 PMCID: PMC11301972 DOI: 10.1186/s12880-024-01379-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2024] [Accepted: 07/25/2024] [Indexed: 08/10/2024] Open
Abstract
BACKGROUND Computed tomography (CT) is widely in clinics and is affected by metal implants. Metal segmentation is crucial for metal artifact correction, and the common threshold method often fails to accurately segment metals. PURPOSE This study aims to segment metal implants in CT images using a diffusion model and further validate it with clinical artifact images and phantom images of known size. METHODS A retrospective study was conducted on 100 patients who received radiation therapy without metal artifacts, and simulated artifact data were generated using publicly available mask data. The study utilized 11,280 slices for training and verification, and 2,820 slices for testing. Metal mask segmentation was performed using DiffSeg, a diffusion model incorporating conditional dynamic coding and a global frequency parser (GFParser). Conditional dynamic coding fuses the current segmentation mask and prior images at multiple scales, while GFParser helps eliminate high-frequency noise in the mask. Clinical artifact images and phantom images are also used for model validation. RESULTS Compared with the ground truth, the accuracy of DiffSeg for metal segmentation of simulated data was 97.89% and that of DSC was 95.45%. The mask shape obtained by threshold segmentation covered the ground truth and DSCs were 82.92% and 84.19% for threshold segmentation based on 2500 HU and 3000 HU. Evaluation metrics and visualization results show that DiffSeg performs better than other classical deep learning networks, especially for clinical CT, artifact data, and phantom data. CONCLUSION DiffSeg efficiently and robustly segments metal masks in artifact data with conditional dynamic coding and GFParser. Future work will involve embedding the metal segmentation model in metal artifact reduction to improve the reduction effect.
Collapse
Affiliation(s)
- Kai Xie
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China
| | - Liugang Gao
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China
| | - Yutao Zhang
- Center for Medical Physics, Nanjing Medical University, Changzhou, 213003, China
- Changzhou Key Laboratory of Medical Physics, Changzhou, 213000, China
| | - Heng Zhang
- Center for Medical Physics, Nanjing Medical University, Changzhou, 213003, China
- Changzhou Key Laboratory of Medical Physics, Changzhou, 213000, China
| | - Jiawei Sun
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China
| | - Tao Lin
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China
| | - Jianfeng Sui
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China
| | - Xinye Ni
- Radiotherapy Department, The Affiliated Changzhou NO.2 People's Hospital of Nanjing Medical University, Changzhou, 213000, China.
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, 213000, China.
- Center for Medical Physics, Nanjing Medical University, Changzhou, 213003, China.
- Changzhou Key Laboratory of Medical Physics, Changzhou, 213000, China.
| |
Collapse
|
7
|
Hu W, Yang S, Guo W, Xiao N, Yang X, Ren X. STC-UNet: renal tumor segmentation based on enhanced feature extraction at different network levels. BMC Med Imaging 2024; 24:179. [PMID: 39030510 PMCID: PMC11264758 DOI: 10.1186/s12880-024-01359-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2024] [Accepted: 07/08/2024] [Indexed: 07/21/2024] Open
Abstract
Renal tumors are one of the common diseases of urology, and precise segmentation of these tumors plays a crucial role in aiding physicians to improve diagnostic accuracy and treatment effectiveness. Nevertheless, inherent challenges associated with renal tumors, such as indistinct boundaries, morphological variations, and uncertainties in size and location, segmenting renal tumors accurately remains a significant challenge in the field of medical image segmentation. With the development of deep learning, substantial achievements have been made in the domain of medical image segmentation. However, existing models lack specificity in extracting features of renal tumors across different network hierarchies, which results in insufficient extraction of renal tumor features and subsequently affects the accuracy of renal tumor segmentation. To address this issue, we propose the Selective Kernel, Vision Transformer, and Coordinate Attention Enhanced U-Net (STC-UNet). This model aims to enhance feature extraction, adapting to the distinctive characteristics of renal tumors across various network levels. Specifically, the Selective Kernel modules are introduced in the shallow layers of the U-Net, where detailed features are more abundant. By selectively employing convolutional kernels of different scales, the model enhances its capability to extract detailed features of renal tumors across multiple scales. Subsequently, in the deeper layers of the network, where feature maps are smaller yet contain rich semantic information, the Vision Transformer modules are integrated in a non-patch manner. These assist the model in capturing long-range contextual information globally. Their non-patch implementation facilitates the capture of fine-grained features, thereby achieving collaborative enhancement of global-local information and ultimately strengthening the model's extraction of semantic features of renal tumors. Finally, in the decoder segment, the Coordinate Attention modules embedding positional information are proposed aiming to enhance the model's feature recovery and tumor region localization capabilities. Our model is validated on the KiTS19 dataset, and experimental results indicate that compared to the baseline model, STC-UNet shows improvements of 1.60%, 2.02%, 2.27%, 1.18%, 1.52%, and 1.35% in IoU, Dice, Accuracy, Precision, Recall, and F1-score, respectively. Furthermore, the experimental results demonstrate that the proposed STC-UNet method surpasses other advanced algorithms in both visual effectiveness and objective evaluation metrics.
Collapse
Affiliation(s)
- Wei Hu
- School of Electrical and Information Engineering of Zhengzhou University, Zhengzhou, China
| | - Shouyi Yang
- School of Electrical and Information Engineering of Zhengzhou University, Zhengzhou, China
| | - Weifeng Guo
- School of Electrical and Information Engineering of Zhengzhou University, Zhengzhou, China.
| | - Na Xiao
- Faculty of Engineering, Huanghe Science and Technology University, Zhengzhou, China
| | - Xiaopeng Yang
- Medical 3D Printing Center of the First Affiliated Hospital of Zhengzhou University, Zhengzhou, China.
| | - Xiangyang Ren
- Medical 3D Printing Center of the First Affiliated Hospital of Zhengzhou University, Zhengzhou, China.
| |
Collapse
|
8
|
Ansari MY, Qaraqe M, Righetti R, Serpedin E, Qaraqe K. Enhancing ECG-based heart age: impact of acquisition parameters and generalization strategies for varying signal morphologies and corruptions. Front Cardiovasc Med 2024; 11:1424585. [PMID: 39027006 PMCID: PMC11254851 DOI: 10.3389/fcvm.2024.1424585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2024] [Accepted: 06/04/2024] [Indexed: 07/20/2024] Open
Abstract
Electrocardiogram (ECG) is a non-invasive approach to capture the overall electrical activity produced by the contraction and relaxation of the cardiac muscles. It has been established in the literature that the difference between ECG-derived age and chronological age represents a general measure of cardiovascular health. Elevated ECG-derived age strongly correlates with cardiovascular conditions (e.g., atherosclerotic cardiovascular disease). However, the neural networks for ECG age estimation are yet to be thoroughly evaluated from the perspective of ECG acquisition parameters. Additionally, deep learning systems for ECG analysis encounter challenges in generalizing across diverse ECG morphologies in various ethnic groups and are susceptible to errors with signals that exhibit random or systematic distortions To address these challenges, we perform a comprehensive empirical study to determine the threshold for the sampling rate and duration of ECG signals while considering their impact on the computational cost of the neural networks. To tackle the concern of ECG waveform variability in different populations, we evaluate the feasibility of utilizing pre-trained and fine-tuned networks to estimate ECG age in different ethnic groups. Additionally, we empirically demonstrate that finetuning is an environmentally sustainable way to train neural networks, and it significantly decreases the ECG instances required (by more than 100 × ) for attaining performance similar to the networks trained from random weight initialization on a complete dataset. Finally, we systematically evaluate augmentation schemes for ECG signals in the context of age estimation and introduce a random cropping scheme that provides best-in-class performance while using shorter-duration ECG signals. The results also show that random cropping enables the networks to perform well with systematic and random ECG signal corruptions.
Collapse
Affiliation(s)
- Mohammed Yusuf Ansari
- Electrical and Computer Engineering, Texas A&M University, College Station, TX, United States
- Electrical and Computer Engineering, Texas A&M University at Qatar, Doha, Qatar
| | - Marwa Qaraqe
- Electrical and Computer Engineering, Texas A&M University at Qatar, Doha, Qatar
- Division of Information and Computing Technology, College of Science and Engineering, Hamad Bin Khalifa University, Qatar Foundation, Doha, Qatar
| | - Raffaella Righetti
- Electrical and Computer Engineering, Texas A&M University, College Station, TX, United States
| | - Erchin Serpedin
- Electrical and Computer Engineering, Texas A&M University, College Station, TX, United States
| | - Khalid Qaraqe
- Electrical and Computer Engineering, Texas A&M University at Qatar, Doha, Qatar
| |
Collapse
|
9
|
Zhang C, Peng J, Wang L, Wang Y, Chen W, Sun MW, Jiang H. A deep learning-powered diagnostic model for acute pancreatitis. BMC Med Imaging 2024; 24:154. [PMID: 38902660 PMCID: PMC11188273 DOI: 10.1186/s12880-024-01339-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2024] [Accepted: 06/17/2024] [Indexed: 06/22/2024] Open
Abstract
BACKGROUND Acute pancreatitis is one of the most common diseases requiring emergency surgery. Rapid and accurate recognition of acute pancreatitis can help improve clinical outcomes. This study aimed to develop a deep learning-powered diagnostic model for acute pancreatitis. MATERIALS AND METHODS In this investigation, we enrolled a cohort of 190 patients with acute pancreatitis who were admitted to Sichuan Provincial People's Hospital between January 2020 and December 2021. Abdominal computed tomography (CT) scans were obtained from both patients with acute pancreatitis and healthy individuals. Our model was constructed using two modules: (1) the acute pancreatitis classifier module; (2) the pancreatitis lesion segmentation module. Each model's performance was assessed based on precision, recall rate, F1-score, Area Under the Curve (AUC), loss rate, frequency-weighted accuracy (fwavacc), and Mean Intersection over Union (MIOU). RESULTS Upon admission, significant variations were observed between patients with mild and severe acute pancreatitis in inflammatory indexes, liver, and kidney function indicators, as well as coagulation parameters. The acute pancreatitis classifier module exhibited commendable diagnostic efficacy, showing an impressive AUC of 0.993 (95%CI: 0.978-0.999) in the test set (comprising healthy examination patients vs. those with acute pancreatitis, P < 0.001) and an AUC of 0.850 (95%CI: 0.790-0.898) in the external validation set (healthy examination patients vs. patients with acute pancreatitis, P < 0.001). Furthermore, the acute pancreatitis lesion segmentation module demonstrated exceptional performance in the validation set. For pancreas segmentation, peripancreatic inflammatory exudation, peripancreatic effusion, and peripancreatic abscess necrosis, the MIOU values were 86.02 (84.52, 87.20), 61.81 (56.25, 64.83), 57.73 (49.90, 68.23), and 66.36 (55.08, 72.12), respectively. These findings underscore the robustness and reliability of the developed models in accurately characterizing and assessing acute pancreatitis. CONCLUSION The diagnostic model for acute pancreatitis, driven by deep learning, exhibits excellent efficacy in accurately evaluating the severity of the condition. TRIAL REGISTRATION This is a retrospective study.
Collapse
Affiliation(s)
- Chi Zhang
- Department of Intensive Care Medicine, The Third People's Hospital of Chengdu, Affiliated Hospital of Southwest Jiaotong University, Chengdu, China
| | - Jin Peng
- Institute for Emergency and Disaster Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China
- Department of Histology and Neuroscience, Sichuan University, Chengdu, China
| | - Lu Wang
- Institute for Emergency and Disaster Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China
- Department of Emergency Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China
- Sichuan Provincial Clinical Research Center for Emergency and Critical Care, School of Medicine, University of Electronic Science and Technology of China, Chengdu, China
| | - Yu Wang
- Sichuan Provincial Clinical Research Center for Emergency and Critical Care, School of Medicine, University of Electronic Science and Technology of China, Chengdu, China
- Department of Clinical Nutrition, Peking Union Medical College Hospital, Beijing, China
| | - Wei Chen
- Department of Clinical Nutrition, Peking Union Medical College Hospital, Beijing, China
| | - Ming-Wei Sun
- Institute for Emergency and Disaster Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China
| | - Hua Jiang
- Institute for Emergency and Disaster Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China.
- Department of Emergency Medicine, School of Medicine, Sichuan Provincial People's Hospital, University of Electronic Science and Technology of China, Chengdu, China.
- Sichuan Provincial Clinical Research Center for Emergency and Critical Care, School of Medicine, University of Electronic Science and Technology of China, Chengdu, China.
| |
Collapse
|
10
|
Dutta K, Laforest R, Luo J, Jha AK, Shoghi KI. Deep learning generation of preclinical positron emission tomography (PET) images from low-count PET with task-based performance assessment. Med Phys 2024; 51:4324-4339. [PMID: 38710222 DOI: 10.1002/mp.17105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2023] [Revised: 04/02/2024] [Accepted: 04/09/2024] [Indexed: 05/08/2024] Open
Abstract
BACKGROUND Preclinical low-count positron emission tomography (LC-PET) imaging offers numerous advantages such as facilitating imaging logistics, enabling longitudinal studies of long- and short-lived isotopes as well as increasing scanner throughput. However, LC-PET is characterized by reduced photon-count levels resulting in low signal-to-noise ratio (SNR), segmentation difficulties, and quantification uncertainties. PURPOSE We developed and evaluated a novel deep-learning (DL) architecture-Attention based Residual-Dilated Net (ARD-Net)-to generate standard-count PET (SC-PET) images from LC-PET images. The performance of the ARD-Net framework was evaluated for numerous low count realizations using fidelity-based qualitative metrics, task-based segmentation, and quantitative metrics. METHOD Patient Derived tumor Xenograft (PDX) with tumors implanted in the mammary fat-pad were subjected to preclinical [18F]-Fluorodeoxyglucose (FDG)-PET/CT imaging. SC-PET images were derived from a 10 min static FDG-PET acquisition, 50 min post administration of FDG, and were resampled to generate four distinct LC-PET realizations corresponding to 10%, 5%, 1.6%, and 0.8% of SC-PET count-level. ARD-Net was trained and optimized using 48 preclinical FDG-PET datasets, while 16 datasets were utilized to assess performance. Further, the performance of ARD-Net was benchmarked against two leading DL-based methods (Residual UNet, RU-Net; and Dilated Network, D-Net) and non-DL methods (Non-Local Means, NLM; and Block Matching 3D Filtering, BM3D). The performance of the framework was evaluated using traditional fidelity-based image quality metrics such as Structural Similarity Index Metric (SSIM) and Normalized Root Mean Square Error (NRMSE), as well as human observer-based tumor segmentation performance (Dice Score and volume bias) and quantitative analysis of Standardized Uptake Value (SUV) measurements. Additionally, radiomics-derived features were utilized as a measure of quality assurance (QA) in comparison to true SC-PET. Finally, a performance ensemble score (EPS) was developed by integrating fidelity-based and task-based metrics. Concordance Correlation Coefficient (CCC) was utilized to determine concordance between measures. The non-parametric Friedman Test with Bonferroni correction was used to compare the performance of ARD-Net against benchmarked methods with significance at adjusted p-value ≤0.01. RESULTS ARD-Net-generated SC-PET images exhibited significantly better (p ≤ 0.01 post Bonferroni correction) overall image fidelity scores in terms of SSIM and NRMSE at majority of photon-count levels compared to benchmarked DL and non-DL methods. In terms of task-based quantitative accuracy evaluated by SUVMean and SUVPeak, ARD-Net exhibited less than 5% median absolute bias for SUVMean compared to true SC-PET and lower degree of variability compared to benchmarked DL and non-DL based methods in generating SC-PET. Additionally, ARD-Net-generated SC-PET images displayed higher degree of concordance to SC-PET images in terms of radiomics features compared to non-DL and other DL approaches. Finally, the ensemble score suggested that ARD-Net exhibited significantly superior performance compared to benchmarked algorithms (p ≤ 0.01 post Bonferroni correction). CONCLUSION ARD-Net provides a robust framework to generate SC-PET from LC-PET images. ARD-Net generated SC-PET images exhibited superior performance compared other DL and non-DL approaches in terms of image-fidelity based metrics, task-based segmentation metrics, and minimal bias in terms of task-based quantification performance for preclinical PET imaging.
Collapse
Affiliation(s)
- Kaushik Dutta
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Missouri, USA
- Imaging Science Program, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
| | - Richard Laforest
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Missouri, USA
- Imaging Science Program, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
| | - Jingqin Luo
- Department of Surgery, Public Health Sciences, Washington University in St Louis, St Louis, Missouri, USA
| | - Abhinav K Jha
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Missouri, USA
- Imaging Science Program, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
- Department of Biomedical Engineering, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
| | - Kooresh I Shoghi
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, Missouri, USA
- Imaging Science Program, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
- Department of Biomedical Engineering, McKelvey School of Engineering, Washington University in St Louis, St Louis, Missouri, USA
| |
Collapse
|
11
|
Wiest IC, Gilbert S, Kather JN. From research to reality: The role of artificial intelligence applications in HCC care. Clin Liver Dis (Hoboken) 2024; 23:e0136. [PMID: 38567094 PMCID: PMC10986906 DOI: 10.1097/cld.0000000000000136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/24/2023] [Accepted: 01/12/2024] [Indexed: 04/04/2024] Open
Affiliation(s)
- Isabella C. Wiest
- Else Kroener Fresenius Center for Digital Health, Medical Faculty Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
- Department of Medicine II, Medical Faculty Mannheim, Heidelberg University, Mannheim, Germany
| | - Stephen Gilbert
- Else Kroener Fresenius Center for Digital Health, Medical Faculty Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
| | - Jakob N. Kather
- Else Kroener Fresenius Center for Digital Health, Medical Faculty Carl Gustav Carus, TUD Dresden University of Technology, Dresden, Germany
- Medical Oncology, National Center for Tumor Diseases (NCT), University Hospital Heidelberg, Heidelberg, Germany
- Department of Medicine I, University Hospital Dresden, Dresden, Germany
| |
Collapse
|
12
|
Wang D, Wang Z, Chen L, Xiao H, Yang B. Cross-Parallel Transformer: Parallel ViT for Medical Image Segmentation. SENSORS (BASEL, SWITZERLAND) 2023; 23:9488. [PMID: 38067861 PMCID: PMC10708613 DOI: 10.3390/s23239488] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 11/22/2023] [Accepted: 11/23/2023] [Indexed: 01/24/2024]
Abstract
Medical image segmentation primarily utilizes a hybrid model consisting of a Convolutional Neural Network and sequential Transformers. The latter leverage multi-head self-attention mechanisms to achieve comprehensive global context modelling. However, despite their success in semantic segmentation, the feature extraction process is inefficient and demands more computational resources, which hinders the network's robustness. To address this issue, this study presents two innovative methods: PTransUNet (PT model) and C-PTransUNet (C-PT model). The C-PT module refines the Vision Transformer by substituting a sequential design with a parallel one. This boosts the feature extraction capabilities of Multi-Head Self-Attention via self-correlated feature attention and channel feature interaction, while also streamlining the Feed-Forward Network to lower computational demands. On the Synapse public dataset, the PT and C-PT models demonstrate improvements in DSC accuracy by 0.87% and 3.25%, respectively, in comparison with the baseline model. As for the parameter count and FLOPs, the PT model aligns with the baseline model. In contrast, the C-PT model shows a decrease in parameter count by 29% and FLOPs by 21.4% relative to the baseline model. The proposed segmentation models in this study exhibit benefits in both accuracy and efficiency.
Collapse
Affiliation(s)
| | | | | | | | - Bo Yang
- College of Engineering and Design, Hunan Normal University, Changsha 410081, China; (D.W.); (Z.W.); (L.C.); (H.X.)
| |
Collapse
|
13
|
Horkaew P, Chansangrat J, Keeratibharat N, Le DC. Recent advances in computerized imaging and its vital roles in liver disease diagnosis, preoperative planning, and interventional liver surgery: A review. World J Gastrointest Surg 2023; 15:2382-2397. [PMID: 38111769 PMCID: PMC10725533 DOI: 10.4240/wjgs.v15.i11.2382] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/26/2023] [Revised: 08/30/2023] [Accepted: 09/27/2023] [Indexed: 11/26/2023] Open
Abstract
The earliest and most accurate detection of the pathological manifestations of hepatic diseases ensures effective treatments and thus positive prognostic outcomes. In clinical settings, screening and determining the extent of a pathology are prominent factors in preparing remedial agents and administering appropriate therapeutic procedures. Moreover, in a patient undergoing liver resection, a realistic preoperative simulation of the subject-specific anatomy and physiology also plays a vital part in conducting initial assessments, making surgical decisions during the procedure, and anticipating postoperative results. Conventionally, various medical imaging modalities, e.g., computed tomography, magnetic resonance imaging, and positron emission tomography, have been employed to assist in these tasks. In fact, several standardized procedures, such as lesion detection and liver segmentation, are also incorporated into prominent commercial software packages. Thus far, most integrated software as a medical device typically involves tedious interactions from the physician, such as manual delineation and empirical adjustments, as per a given patient. With the rapid progress in digital health approaches, especially medical image analysis, a wide range of computer algorithms have been proposed to facilitate those procedures. They include pattern recognition of a liver, its periphery, and lesion, as well as pre- and postoperative simulations. Prior to clinical adoption, however, software must conform to regulatory requirements set by the governing agency, for instance, valid clinical association and analytical and clinical validation. Therefore, this paper provides a detailed account and discussion of the state-of-the-art methods for liver image analyses, visualization, and simulation in the literature. Emphasis is placed upon their concepts, algorithmic classifications, merits, limitations, clinical considerations, and future research trends.
Collapse
Affiliation(s)
- Paramate Horkaew
- School of Computer Engineering, Suranaree University of Technology, Nakhon Ratchasima 30000, Thailand
| | - Jirapa Chansangrat
- School of Radiology, Institute of Medicine, Suranaree University of Technology, Nakhon Ratchasima 30000, Thailand
| | - Nattawut Keeratibharat
- School of Surgery, Institute of Medicine, Suranaree University of Technology, Nakhon Ratchasima 30000, Thailand
| | - Doan Cong Le
- Faculty of Information Technology, An Giang University, Vietnam National University (Ho Chi Minh City), An Giang 90000, Vietnam
| |
Collapse
|
14
|
Ansari Y, Mourad O, Qaraqe K, Serpedin E. Deep learning for ECG Arrhythmia detection and classification: an overview of progress for period 2017-2023. Front Physiol 2023; 14:1246746. [PMID: 37791347 PMCID: PMC10542398 DOI: 10.3389/fphys.2023.1246746] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2023] [Accepted: 08/28/2023] [Indexed: 10/05/2023] Open
Abstract
Cardiovascular diseases are a leading cause of mortality globally. Electrocardiography (ECG) still represents the benchmark approach for identifying cardiac irregularities. Automatic detection of abnormalities from the ECG can aid in the early detection, diagnosis, and prevention of cardiovascular diseases. Deep Learning (DL) architectures have been successfully employed for arrhythmia detection and classification and offered superior performance to traditional shallow Machine Learning (ML) approaches. This survey categorizes and compares the DL architectures used in ECG arrhythmia detection from 2017-2023 that have exhibited superior performance. Different DL models such as Convolutional Neural Networks (CNNs), Multilayer Perceptrons (MLPs), Transformers, and Recurrent Neural Networks (RNNs) are reviewed, and a summary of their effectiveness is provided. This survey provides a comprehensive roadmap to expedite the acclimation process for emerging researchers willing to develop efficient algorithms for detecting ECG anomalies using DL models. Our tailored guidelines bridge the knowledge gap allowing newcomers to align smoothly with the prevailing research trends in ECG arrhythmia detection. We shed light on potential areas for future research and refinement in model development and optimization, intending to stimulate advancement in ECG arrhythmia detection and classification.
Collapse
Affiliation(s)
- Yaqoob Ansari
- ECEN Program, Texas A&M University at Qatar, Doha, Qatar
| | | | - Khalid Qaraqe
- ECEN Program, Texas A&M University at Qatar, Doha, Qatar
| | - Erchin Serpedin
- ECEN Department, Texas A&M University, College Station, TX, United States
| |
Collapse
|
15
|
Rai P, Ansari MY, Warfa M, Al-Hamar H, Abinahed J, Barah A, Dakua SP, Balakrishnan S. Efficacy of fusion imaging for immediate post-ablation assessment of malignant liver neoplasms: A systematic review. Cancer Med 2023. [PMID: 37191030 DOI: 10.1002/cam4.6089] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 04/27/2023] [Accepted: 05/05/2023] [Indexed: 05/17/2023] Open
Abstract
BACKGROUND Percutaneous thermal ablation has become the preferred therapeutic treatment option for liver cancers that cannot be resected. Since ablative zone tissue changes over time, it becomes challenging to determine therapy effectiveness over an extended period. Thus, an immediate post-procedural evaluation of the ablation zone is crucial, as it could influence the need for a second-look treatment or follow-up plan. Assessing treatment response immediately after ablation is essential to attain favorable outcomes. This study examines the efficacy of image fusion strategies immediately post-ablation in liver neoplasms to determine therapeutic response. METHODOLOGY A comprehensive systematic search using PRISMA methodology was conducted using EMBASE, MEDLINE (via PUBMED), and Cochrane Library Central Registry electronic databases to identify articles that assessed the immediate post-ablation response in malignant hepatic tumors with fusion imaging (FI) systems. The data were retrieved on relevant clinical characteristics, including population demographics, pre-intervention clinical history, lesion characteristics, and intervention type. For the outcome metrics, variables such as average fusion time, intervention metrics, technical success rate, ablative safety margin, supplementary ablation rate, technical efficacy rate, LTP rates, and reported complications were extracted. RESULTS Twenty-two studies were included for review after fulfilling the study eligibility criteria. FI's immediate technical success rate ranged from 81.3% to 100% in 17/22 studies. In 16/22 studies, the ablative safety margin was assessed immediately after ablation. Supplementary ablation was performed in 9 studies following immediate evaluation by FI. In 15/22 studies, the technical effectiveness rates during the first follow-up varied from 89.3% to 100%. CONCLUSION Based on the studies included, we found that FI can accurately determine the immediate therapeutic response in liver cancer ablation image fusion and could be a feasible intraprocedural tool for determining short-term post-ablation outcomes in unresectable liver neoplasms. There are some technical challenges that limit the widespread adoption of FI techniques. Large-scale randomized trials are warranted to improve on existing protocols. Future research should emphasize improving FI's technological capabilities and clinical applicability to a broader range of tumor types and ablation procedures.
Collapse
Affiliation(s)
- Pragati Rai
- Department of Surgery, Hamad Medical Corporation, Doha, Qatar
| | | | - Mohammed Warfa
- Department of Clinical Imaging, Wake Forest Baptist Medical Center, Winston-Salem, North Carolina, USA
| | - Hammad Al-Hamar
- Department of Clinical Imaging, Hamad Medical Corporation, Doha, Qatar
| | - Julien Abinahed
- Department of Surgery, Hamad Medical Corporation, Doha, Qatar
| | - Ali Barah
- Department of Clinical Imaging, Hamad Medical Corporation, Doha, Qatar
| | | | | |
Collapse
|
16
|
Ansari MY, Abdalla A, Ansari MY, Ansari MI, Malluhi B, Mohanty S, Mishra S, Singh SS, Abinahed J, Al-Ansari A, Balakrishnan S, Dakua SP. Correction: Practical utility of liver segmentation methods in clinical surgeries and interventions. BMC Med Imaging 2022; 22:141. [PMID: 35941585 PMCID: PMC9361601 DOI: 10.1186/s12880-022-00869-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022] Open
|