1
|
Jain S, Li X, Xu M. Knowledge Transfer from Macro-world to Micro-world: Enhancing 3D Cryo-ET Classification through Fine-Tuning Video-based Deep Models. Bioinformatics 2024; 40:btae368. [PMID: 38889274 PMCID: PMC11269433 DOI: 10.1093/bioinformatics/btae368] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2024] [Revised: 04/30/2024] [Accepted: 06/11/2024] [Indexed: 06/20/2024] Open
Abstract
MOTIVATION Deep learning models have achieved remarkable success in a wide range of natural-world tasks, such as vision, language, and speech recognition. These accomplishments are largely attributed to the availability of open-source large-scale datasets. More importantly, pre-trained foundational modellearnings exhibit a surprising degree of transferability to downstream tasks, enabling efficient learning even with limited training examples. However, the application of such natural-domain models to the domain of tiny Cryo-Electron Tomography (Cryo-ET) images has been a relatively unexplored frontier. This research is motivated by the intuition that 3D Cryo-ET voxel data can be conceptually viewed as a sequence of progressively evolving video frames. RESULTS Leveraging the above insight, we propose a novel approach that involves the utilization of 3D models pre-trained on large-scale video datasets to enhance Cryo-ET subtomogram classification. Our experiments, conducted on both simulated and real Cryo-ET datasets, reveal compelling results. The use of video initialization not only demonstrates improvements in classification accuracy but also substantially reduces training costs. Further analyses provide additional evidence of the value of video initialization in enhancing subtomogram feature extraction. Additionally, we observe that video initialization yields similar positive effects when applied to medical 3D classification tasks, underscoring the potential of cross-domain knowledge transfer from video-based models to advance the state-of-the-art in a wide range of biological and medical data types. AVAILABILITY AND IMPLEMENTATION https://github.com/xulabs/aitom.
Collapse
Affiliation(s)
- Sabhay Jain
- Electrical Engineering Department, Indian Institute of Technology Kanpur, Kanpur, Uttar Pradesh, 208016, India
| | - Xingjian Li
- Ray and Stephanie Lane Computational Biology Department, Carnegie Mellon University, Pittsburgh, Pennsylvania, 15213, United States
| | - Min Xu
- Ray and Stephanie Lane Computational Biology Department, Carnegie Mellon University, Pittsburgh, Pennsylvania, 15213, United States
| |
Collapse
|
2
|
Diao S, Chen P, Showkatian E, Bandyopadhyay R, Rojas FR, Zhu B, Hong L, Aminu M, Saad MB, Salehjahromi M, Muneer A, Sujit SJ, Behrens C, Gibbons DL, Heymach JV, Kalhor N, Wistuba II, Solis Soto LM, Zhang J, Qin W, Wu J. Automated Cellular-Level Dual Global Fusion of Whole-Slide Imaging for Lung Adenocarcinoma Prognosis. Cancers (Basel) 2023; 15:4824. [PMID: 37835518 PMCID: PMC10571722 DOI: 10.3390/cancers15194824] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 09/24/2023] [Accepted: 09/27/2023] [Indexed: 10/15/2023] Open
Abstract
Histopathologic whole-slide images (WSI) are generally considered the gold standard for cancer diagnosis and prognosis. Survival prediction based on WSI has recently attracted substantial attention. Nevertheless, it remains a central challenge owing to the inherent difficulties of predicting patient prognosis and effectively extracting informative survival-specific representations from WSI with highly compounded gigapixels. In this study, we present a fully automated cellular-level dual global fusion pipeline for survival prediction. Specifically, the proposed method first describes the composition of different cell populations on WSI. Then, it generates dimension-reduced WSI-embedded maps, allowing for efficient investigation of the tumor microenvironment. In addition, we introduce a novel dual global fusion network to incorporate global and inter-patch features of cell distribution, which enables the sufficient fusion of different types and locations of cells. We further validate the proposed pipeline using The Cancer Genome Atlas lung adenocarcinoma dataset. Our model achieves a C-index of 0.675 (±0.05) in the five-fold cross-validation setting and surpasses comparable methods. Further, we extensively analyze embedded map features and survival probabilities. These experimental results manifest the potential of our proposed pipeline for applications using WSI in lung adenocarcinoma and other malignancies.
Collapse
Affiliation(s)
- Songhui Diao
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
- Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055, China
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Pingjun Chen
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Eman Showkatian
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Rukhmini Bandyopadhyay
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Frank R. Rojas
- Department of Translational Molecular Pathology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Bo Zhu
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Lingzhi Hong
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Muhammad Aminu
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Maliazurina B. Saad
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Morteza Salehjahromi
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Amgad Muneer
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Sheeba J. Sujit
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Carmen Behrens
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Don L. Gibbons
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - John V. Heymach
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Neda Kalhor
- Department of Pathology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Ignacio I. Wistuba
- Department of Translational Molecular Pathology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Luisa M. Solis Soto
- Department of Translational Molecular Pathology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Jianjun Zhang
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
- Department of Genomic Medicine, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Wenjian Qin
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Jia Wu
- Department of Imaging Physics, Division of Diagnostic Imaging, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
- Department of Thoracic/Head and Neck Medical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| |
Collapse
|
3
|
Park S, Kim JH, Cha YK, Chung MJ, Woo JH, Park S. Application of Machine Learning Algorithm in Predicting Axillary Lymph Node Metastasis from Breast Cancer on Preoperative Chest CT. Diagnostics (Basel) 2023; 13:2953. [PMID: 37761320 PMCID: PMC10528867 DOI: 10.3390/diagnostics13182953] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 09/05/2023] [Accepted: 09/13/2023] [Indexed: 09/29/2023] Open
Abstract
Axillary lymph node (ALN) status is one of the most critical prognostic factors in patients with breast cancer. However, ALN evaluation with contrast-enhanced CT (CECT) has been challenging. Machine learning (ML) is known to show excellent performance in image recognition tasks. The purpose of our study was to evaluate the performance of the ML algorithm for predicting ALN metastasis by combining preoperative CECT features of both ALN and primary tumor. This was a retrospective single-institutional study of a total of 266 patients with breast cancer who underwent preoperative chest CECT. Random forest (RF), extreme gradient boosting (XGBoost), and neural network (NN) algorithms were used. Statistical analysis and recursive feature elimination (RFE) were adopted as feature selection for ML. The best ML-based ALN prediction model for breast cancer was NN with RFE, which achieved an AUROC of 0.76 ± 0.11 and an accuracy of 0.74 ± 0.12. By comparing NN with RFE model performance with and without ALN features from CECT, NN with RFE model with ALN features showed better performance at all performance evaluations, which indicated the effect of ALN features. Through our study, we were able to demonstrate that the ML algorithm could effectively predict the final diagnosis of ALN metastases from CECT images of the primary tumor and ALN. This suggests that ML has the potential to differentiate between benign and malignant ALNs.
Collapse
Affiliation(s)
- Soyoung Park
- Department of Health Sciences and Technology, SAIHST, Sungkyunkwan University, Seoul 06351, Republic of Korea; (S.P.); (S.P.)
| | - Jong Hee Kim
- Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.H.K.); (J.H.W.)
| | - Yoon Ki Cha
- Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.H.K.); (J.H.W.)
| | - Myung Jin Chung
- Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.H.K.); (J.H.W.)
| | - Jung Han Woo
- Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.H.K.); (J.H.W.)
| | - Subin Park
- Department of Health Sciences and Technology, SAIHST, Sungkyunkwan University, Seoul 06351, Republic of Korea; (S.P.); (S.P.)
| |
Collapse
|
4
|
Sauter D, Lodde G, Nensa F, Schadendorf D, Livingstone E, Kukuk M. Deep learning in computational dermatopathology of melanoma: A technical systematic literature review. Comput Biol Med 2023; 163:107083. [PMID: 37315382 DOI: 10.1016/j.compbiomed.2023.107083] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Revised: 05/10/2023] [Accepted: 05/27/2023] [Indexed: 06/16/2023]
Abstract
Deep learning (DL) has become one of the major approaches in computational dermatopathology, evidenced by a significant increase in this topic in the current literature. We aim to provide a structured and comprehensive overview of peer-reviewed publications on DL applied to dermatopathology focused on melanoma. In comparison to well-published DL methods on non-medical images (e.g., classification on ImageNet), this field of application comprises a specific set of challenges, such as staining artifacts, large gigapixel images, and various magnification levels. Thus, we are particularly interested in the pathology-specific technical state-of-the-art. We also aim to summarize the best performances achieved thus far with respect to accuracy, along with an overview of self-reported limitations. Accordingly, we conducted a systematic literature review of peer-reviewed journal and conference articles published between 2012 and 2022 in the databases ACM Digital Library, Embase, IEEE Xplore, PubMed, and Scopus, expanded by forward and backward searches to identify 495 potentially eligible studies. After screening for relevance and quality, a total of 54 studies were included. We qualitatively summarized and analyzed these studies from technical, problem-oriented, and task-oriented perspectives. Our findings suggest that the technical aspects of DL for histopathology in melanoma can be further improved. The DL methodology was adopted later in this field, and still lacks the wider adoption of DL methods already shown to be effective for other applications. We also discuss upcoming trends toward ImageNet-based feature extraction and larger models. While DL has achieved human-competitive accuracy in routine pathological tasks, its performance on advanced tasks is still inferior to wet-lab testing (for example). Finally, we discuss the challenges impeding the translation of DL methods to clinical practice and provide insight into future research directions.
Collapse
Affiliation(s)
- Daniel Sauter
- Department of Computer Science, Fachhochschule Dortmund, 44227 Dortmund, Germany.
| | - Georg Lodde
- Department of Dermatology, University Hospital Essen, 45147 Essen, Germany
| | - Felix Nensa
- Institute for AI in Medicine (IKIM), University Hospital Essen, 45131 Essen, Germany; Institute of Diagnostic and Interventional Radiology and Neuroradiology, University Hospital Essen, 45147 Essen, Germany
| | - Dirk Schadendorf
- Department of Dermatology, University Hospital Essen, 45147 Essen, Germany
| | | | - Markus Kukuk
- Department of Computer Science, Fachhochschule Dortmund, 44227 Dortmund, Germany
| |
Collapse
|
5
|
Kalpana B, Reshmy A, Senthil Pandi S, Dhanasekaran S. OESV-KRF: Optimal ensemble support vector kernel random forest based early detection and classification of skin diseases. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104779] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023]
|
6
|
You H, Yu L, Tian S, Cai W. A stereo spatial decoupling network for medical image classification. COMPLEX INTELL SYST 2023; 9:1-10. [PMID: 37361963 PMCID: PMC10107597 DOI: 10.1007/s40747-023-01049-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2022] [Accepted: 03/09/2023] [Indexed: 06/28/2023]
Abstract
Deep convolutional neural network (CNN) has made great progress in medical image classification. However, it is difficult to establish effective spatial associations, and always extracts similar low-level features, resulting in redundancy of information. To solve these limitations, we propose a stereo spatial discoupling network (TSDNets), which can leverage the multi-dimensional spatial details of medical images. Then, we use an attention mechanism to progressively extract the most discriminative features from three directions: horizontal, vertical, and depth. Moreover, a cross feature screening strategy is used to divide the original feature maps into three levels: important, secondary and redundant. Specifically, we design a cross feature screening module (CFSM) and a semantic guided decoupling module (SGDM) to model multi-dimension spatial relationships, thereby enhancing the feature representation capabilities. The extensive experiments conducted on multiple open source baseline datasets demonstrate that our TSDNets outperforms previous state-of-the-art models.
Collapse
Affiliation(s)
- Hongfeng You
- School of Information Science and Engineering, Xinjiang University, Urumqi, 830000 China
| | - Long Yu
- Network Center, Xinjiang University, Urumqi, 830000 China
| | - Shengwei Tian
- Software College, Xinjiang University, Urumqi, 830000 China
| | - Weiwei Cai
- School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi, 214122 China
| |
Collapse
|
7
|
Zhang M, Xue M, Li S, Zou Y, Zhu Q. Fusion deep learning approach combining diffuse optical tomography and ultrasound for improving breast cancer classification. BIOMEDICAL OPTICS EXPRESS 2023; 14:1636-1646. [PMID: 37078047 PMCID: PMC10110311 DOI: 10.1364/boe.486292] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2023] [Revised: 02/25/2023] [Accepted: 03/04/2023] [Indexed: 05/03/2023]
Abstract
Diffuse optical tomography (DOT) is a promising technique that provides functional information related to tumor angiogenesis. However, reconstructing the DOT function map of a breast lesion is an ill-posed and underdetermined inverse process. A co-registered ultrasound (US) system that provides structural information about the breast lesion can improve the localization and accuracy of DOT reconstruction. Additionally, the well-known US characteristics of benign and malignant breast lesions can further improve cancer diagnosis based on DOT alone. Inspired by a fusion model deep learning approach, we combined US features extracted by a modified VGG-11 network with images reconstructed from a DOT deep learning auto-encoder-based model to form a new neural network for breast cancer diagnosis. The combined neural network model was trained with simulation data and fine-tuned with clinical data: it achieved an AUC of 0.931 (95% CI: 0.919-0.943), superior to those achieved using US images alone (0.860) or DOT images alone (0.842).
Collapse
Affiliation(s)
- Menghao Zhang
- Electrical and System Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
| | - Minghao Xue
- Biomedical Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
| | - Shuying Li
- Biomedical Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
| | - Yun Zou
- Biomedical Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
| | - Quing Zhu
- Electrical and System Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
- Biomedical Engineering Department, Washington University in St. Louis, 1 Brooking Dr, St. Louis, MO 63130, USA
| |
Collapse
|
8
|
Ngombu S, Binol H, Gurcan MN, Moberly AC. Advances in Artificial Intelligence to Diagnose Otitis Media: State of the Art Review. Otolaryngol Head Neck Surg 2023; 168:635-642. [PMID: 35290142 DOI: 10.1177/01945998221083502] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2021] [Accepted: 02/09/2022] [Indexed: 11/15/2022]
Abstract
OBJECTIVE Otitis media (OM) is a model disease for developing, validating, and implementing artificial intelligence (AI) techniques. We aim to review the state of the art applications of AI used to diagnose OM in pediatric and adult populations. DATA SOURCES Several comprehensive databases were searched to identify all articles that applied AI technologies to diagnose OM. REVIEW METHODS Relevant articles from January 2010 through May 2021 were identified by title and abstract. Articles were excluded if they did not discuss AI in conjunction with diagnosing OM. References of included studies and relevant review articles were cross-referenced to identify any additional studies. CONCLUSION Title and abstract screening resulted in full-text retrieval of 40 articles that met initial screening parameters. Of this total, secondary review articles (n = 7) and commentary-based articles (n = 2) were removed, as were articles that did not specifically discuss AI and OM diagnosis (n = 5), leaving 25 articles for review. Applications of AI technologies specific to diagnosing OM included machine learning and natural language processing (n = 23) and prototype approaches (n = 2). IMPLICATIONS FOR PRACTICE This review emphasizes the utility of AI techniques to automate and aid in diagnosing OM. Although these techniques are still in the development and testing stages, AI has the potential to improve the practice of otolaryngologists and primary care clinicians by increasing the efficiency and accuracy of diagnoses.
Collapse
Affiliation(s)
- Stephany Ngombu
- Department of Otolaryngology-Head and Neck Surgery, Wexner Medical Center at The Ohio State University, Columbus, Ohio, USA
| | - Hamidullah Binol
- Center for Biomedical Informatics, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
| | - Metin N Gurcan
- Center for Biomedical Informatics, Wake Forest School of Medicine, Winston-Salem, North Carolina, USA
| | - Aaron C Moberly
- Department of Otolaryngology-Head and Neck Surgery, Wexner Medical Center at The Ohio State University, Columbus, Ohio, USA
| |
Collapse
|
9
|
Kurtz MA, Yang R, Elapolu MSR, Wessinger AC, Nelson W, Alaniz K, Rai R, Gilbert JL. Predicting Corrosion Damage in the Human Body Using Artificial Intelligence: In Vitro Progress and Future Applications. Orthop Clin North Am 2023; 54:169-192. [PMID: 36894290 DOI: 10.1016/j.ocl.2022.11.004] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/11/2023]
Abstract
Artificial intelligence (AI) is used in the clinic to improve patient care. While the successes illustrate AI's impact, few studies have led to improved clinical outcomes. In this review, we focus on how AI models implemented in nonorthopedic fields of corrosion science may apply to the study of orthopedic alloys. We first define and introduce fundamental AI concepts and models, as well as physiologically relevant corrosion damage modes. We then systematically review the corrosion/AI literature. Finally, we identify several AI models that may be implemented to study fretting, crevice, and pitting corrosion of titanium and cobalt chrome alloys.
Collapse
Affiliation(s)
- Michael A Kurtz
- Department of Bioengineering, Clemson University, Clemson, SC, USA; The Clemson University-Medical University of South Carolina Bioengineering Program, 68 President Street, Charleston, SC 29425, USA
| | - Ruoyu Yang
- Department of Automotive Engineering, Clemson University, 4 Research Drive, Greenville, SC 29607, USA
| | - Mohan S R Elapolu
- Department of Automotive Engineering, Clemson University, 4 Research Drive, Greenville, SC 29607, USA
| | - Audrey C Wessinger
- Department of Bioengineering, Clemson University, Clemson, SC, USA; The Clemson University-Medical University of South Carolina Bioengineering Program, 68 President Street, Charleston, SC 29425, USA
| | - William Nelson
- Department of Bioengineering, Clemson University, Clemson, SC, USA; The Clemson University-Medical University of South Carolina Bioengineering Program, 68 President Street, Charleston, SC 29425, USA
| | - Kazzandra Alaniz
- Department of Bioengineering, Clemson University, Clemson, SC, USA; The Clemson University-Medical University of South Carolina Bioengineering Program, 68 President Street, Charleston, SC 29425, USA
| | - Rahul Rai
- Department of Automotive Engineering, Clemson University, 4 Research Drive, Greenville, SC 29607, USA
| | - Jeremy L Gilbert
- Department of Bioengineering, Clemson University, Clemson, SC, USA; The Clemson University-Medical University of South Carolina Bioengineering Program, 68 President Street, Charleston, SC 29425, USA.
| |
Collapse
|
10
|
Yousef RN, Khalil AT, Samra AS, Ata MM. Model-based and model-free deep features fusion for high performed human gait recognition. THE JOURNAL OF SUPERCOMPUTING 2023; 79:1-38. [PMID: 37359324 PMCID: PMC10024915 DOI: 10.1007/s11227-023-05156-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Accepted: 03/01/2023] [Indexed: 06/28/2023]
Abstract
In the last decade, the need for a non-contact biometric model for recognizing candidates has increased, especially after the pandemic of COVID-19 appeared and spread worldwide. This paper presents a novel deep convolutional neural network (CNN) model that guarantees quick, safe, and precise human authentication via their poses and walking style. The concatenated fusion between the proposed CNN and a fully connected model has been formulated, utilized, and tested. The proposed CNN extracts the human features from two main sources: (1) human silhouette images according to model-free and (2) human joints, limbs, and static joint distances according to a model-based via a novel, fully connected deep-layer structure. The most commonly used dataset, CASIA gait families, has been utilized and tested. Numerous performance metrics have been evaluated to measure the system quality, including accuracy, specificity, sensitivity, false negative rate, and training time. Experimental results reveal that the proposed model can enhance recognition performance in a superior manner compared with the latest state-of-the-art studies. Moreover, the suggested system introduces a robust real-time authentication with any covariate conditions, scoring 99.8% and 99.6% accuracy in identifying casia (B) and casia (A) datasets, respectively.
Collapse
Affiliation(s)
- Reem N. Yousef
- Delta Higher Institute for Engineering and Technology, Mansoura, Egypt
| | - Abeer T. Khalil
- Electronics and Communications Engineering Department, Faculty of Engineering, Mansoura University, Mansoura, Egypt
| | - Ahmed S. Samra
- Electronics and Communications Engineering Department, Faculty of Engineering, Mansoura University, Mansoura, Egypt
| | - Mohamed Maher Ata
- Department of Communications and Electronics Engineering, MISR Higher Institute for Engineering and Technology, Mansoura, Egypt
| |
Collapse
|
11
|
Anisha A, Jiji G, Ajith Bosco Raj T. Deep feature fusion and optimized feature selection based ensemble classification of liver lesions. THE IMAGING SCIENCE JOURNAL 2023. [DOI: 10.1080/13682199.2023.2185430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/11/2023]
Affiliation(s)
- A. Anisha
- Department of Computer Science and Engineering, St. Xavier’s Catholic College of Engineering, Nagercoil, Tamil Nadu, India
| | - G. Jiji
- Department of Electronics and Communication Engineering, Lord Jegannath College of Engineering and Technology, Nagercoil, Tamil Nadu, India
| | - T. Ajith Bosco Raj
- Department of Electronics and Communication Engineering, PSN College of Engineering and Technology, Tirunelveli, Tamil Nadu, India
| |
Collapse
|
12
|
Gheisari M, Ebrahimzadeh F, Rahimi M, Moazzamigodarzi M, Liu Y, Dutta Pramanik PK, Heravi MA, Mehbodniya A, Ghaderzadeh M, Feylizadeh MR, Kosari S. Deep learning: Applications, architectures, models, tools, and frameworks: A comprehensive survey. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY 2023. [DOI: 10.1049/cit2.12180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/25/2023] Open
Affiliation(s)
- Mehdi Gheisari
- School of Computer Science and Technology Harbin Institute of Technology (Shenzhen) Shenzhen China
- Department of Cognitive Computing, Institute of Computer Science and Engineering, Saveetha School of Engineering Saveetha Institute of Medical and Technical Sciences Chennai India
- Department of Computer Science Islamic Azad University Tehran Iran
| | | | - Mohamadtaghi Rahimi
- Department of Mathematics and Statistics Iran University of Science and Technology Tehran Iran
| | | | - Yang Liu
- School of Computer Science and Technology Harbin Institute of Technology (Shenzhen) Shenzhen China
- Peng Cheng Laboratory Shenzhen China
| | | | | | - Abolfazl Mehbodniya
- Department of Electronics and Communications Engineering Kuwait College of Science and Technology Doha District Kuwait
| | - Mustafa Ghaderzadeh
- Department of Artificial Intelligence Smart University of Medical Sciences Tehran Iran
| | | | - Saeed Kosari
- Institute of Computing Science and Technology, Guangzhou University Guangzhou China
| |
Collapse
|
13
|
Araújo ALD, da Silva VM, Kudo MS, de Souza ESC, Saldivia-Siracusa C, Giraldo-Roldán D, Lopes MA, Vargas PA, Khurram SA, Pearson AT, Kowalski LP, de Carvalho ACPDLF, Santos-Silva AR, Moraes MC. Machine learning concepts applied to oral pathology and oral medicine: A convolutional neural networks' approach. J Oral Pathol Med 2023; 52:109-118. [PMID: 36599081 DOI: 10.1111/jop.13397] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Revised: 12/05/2022] [Accepted: 12/15/2022] [Indexed: 01/06/2023]
Abstract
INTRODUCTION Artificial intelligence models and networks can learn and process dense information in a short time, leading to an efficient, objective, and accurate clinical and histopathological analysis, which can be useful to improve treatment modalities and prognostic outcomes. This paper targets oral pathologists, oral medicinists, and head and neck surgeons to provide them with a theoretical and conceptual foundation of artificial intelligence-based diagnostic approaches, with a special focus on convolutional neural networks, the state-of-the-art in artificial intelligence and deep learning. METHODS The authors conducted a literature review, and the convolutional neural network's conceptual foundations and functionality were illustrated based on a unique interdisciplinary point of view. CONCLUSION The development of artificial intelligence-based models and computer vision methods for pattern recognition in clinical and histopathological image analysis of head and neck cancer has the potential to aid diagnosis and prognostic prediction.
Collapse
Affiliation(s)
- Anna Luíza Damaceno Araújo
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil.,Head and Neck Surgery Department and LIM 28, University of São Paulo Medical School, São Paulo, São Paulo, Brazil
| | - Viviane Mariano da Silva
- Institute of Science and Technology, Federal University of São Paulo (ICT-Unifesp), São José dos Campos, São Paulo, Brazil
| | - Maíra Suzuka Kudo
- Institute of Science and Technology, Federal University of São Paulo (ICT-Unifesp), São José dos Campos, São Paulo, Brazil
| | | | - Cristina Saldivia-Siracusa
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil
| | - Daniela Giraldo-Roldán
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil
| | - Marcio Ajudarte Lopes
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil
| | - Pablo Agustin Vargas
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil
| | - Syed Ali Khurram
- Unit of Oral and Maxillofacial Pathology, School of Clinical Dentistry, University of Sheffield, Sheffield, UK
| | - Alexander T Pearson
- Section of Hemathology/Oncology, Department of Medicine, University of Chicago, Chicago, Illinois, USA.,University of Chicago Comprehensive Cancer Center, Chicago, Illinois, USA
| | - Luiz Paulo Kowalski
- Head and Neck Surgery Department and LIM 28, University of São Paulo Medical School, São Paulo, São Paulo, Brazil.,Department of Head and Neck Surgery and Otorhinolaryngology, A.C. Camargo Cancer Center, São Paulo, São Paulo, Brazil
| | | | - Alan Roger Santos-Silva
- Oral Diagnosis Department, Piracicaba Dental School, University of Campinas (FOP-UNICAMP), Piracicaba, São Paulo, Brazil
| | - Matheus Cardoso Moraes
- Institute of Science and Technology, Federal University of São Paulo (ICT-Unifesp), São José dos Campos, São Paulo, Brazil
| |
Collapse
|
14
|
Kakehbaraei S, Arvanaghi R, Seyedarabi H, Esmaeili F, Zenouz AT. 3D tooth segmentation in cone-beam computed tomography images using distance transform. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
15
|
Wang S, Xu X, Du H, Chen Y, Mei W. Attention feature fusion methodology with additional constraint for ovarian lesion diagnosis on magnetic resonance images. Med Phys 2023; 50:297-310. [PMID: 35975618 DOI: 10.1002/mp.15937] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2021] [Revised: 06/25/2022] [Accepted: 07/24/2022] [Indexed: 01/25/2023] Open
Abstract
PURPOSE It is challenging for radiologists and gynecologists to identify the type of ovarian lesions by reading magnetic resonance (MR) images. Recently developed convolutional neural networks (CNNs) have made great progress in computer vision, but their architectures still need modification if they are used in processing medical images. This study aims to improve the feature extraction capability of CNNs, thus promoting the diagnostic performance in discriminating between benign and malignant ovarian lesions. METHODS We introduce a feature fusion architecture and insert the attention models in the neural network. The features extracted from different middle layers are integrated with reoptimized spatial and channel weights. We add a loss function to constrain the additional probability vector generated from the integrated features, thus guiding the middle layers to emphasize useful information. We analyzed 159 lesions imaged by dynamic contrast-enhanced MR imaging (DCE-MRI), including 73 benign lesions and 86 malignant lesions. Senior radiologists selected and labeled the tumor regions based on the pathology reports. Then, the tumor regions were cropped into 7494 nonoverlapping image patches for training and testing. The type of a single tumor was determined by the average probability scores of the image patches belonging to it. RESULTS We implemented fivefold cross-validation to characterize our proposed method, and the distribution of performance matrics was reported. For all the test image patches, the average accuracy of our method is 70.5% with an average area under the curve (AUC) of 0.785, while the baseline is 69.4% and 0.773, and for the diagnosis of single tumors, our model achieved an average accuracy of 82.4% and average AUC of 0.916, which were better than the baseline (81.8% and 0.899). Moreover, we evaluated the performance of our proposed method utilizing different CNN backbones and different attention mechanisms. CONCLUSIONS The texture features extracted from different middle layers are crucial for ovarian lesion diagnosis. Our proposed method can enhance the feature extraction capabilities of different layers of the network, thereby improving diagnostic performance.
Collapse
Affiliation(s)
- Shuai Wang
- School of Information and Electronics, Beijing Institute of Technology, Beijing, China
| | - Xiaojuan Xu
- Department of Diagnostic Imaging, National Cancer Center, National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking, Union Medical College, Beijing, China
| | - Huiqian Du
- School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing, China
| | - Yan Chen
- Department of Diagnostic Imaging, National Cancer Center, National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking, Union Medical College, Beijing, China
| | - Wenbo Mei
- School of Information and Electronics, Beijing Institute of Technology, Beijing, China
| |
Collapse
|
16
|
Chicco D, Shiradkar R. Ten quick tips for computational analysis of medical images. PLoS Comput Biol 2023; 19:e1010778. [PMID: 36602952 PMCID: PMC9815662 DOI: 10.1371/journal.pcbi.1010778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023] Open
Abstract
Medical imaging is a great asset for modern medicine, since it allows physicians to spatially interrogate a disease site, resulting in precise intervention for diagnosis and treatment, and to observe particular aspect of patients' conditions that otherwise would not be noticeable. Computational analysis of medical images, moreover, can allow the discovery of disease patterns and correlations among cohorts of patients with the same disease, thus suggesting common causes or providing useful information for better therapies and cures. Machine learning and deep learning applied to medical images, in particular, have produced new, unprecedented results that can pave the way to advanced frontiers of medical discoveries. While computational analysis of medical images has become easier, however, the possibility to make mistakes or generate inflated or misleading results has become easier, too, hindering reproducibility and deployment. In this article, we provide ten quick tips to perform computational analysis of medical images avoiding common mistakes and pitfalls that we noticed in multiple studies in the past. We believe our ten guidelines, if taken into practice, can help the computational-medical imaging community to perform better scientific research that eventually can have a positive impact on the lives of patients worldwide.
Collapse
Affiliation(s)
- Davide Chicco
- Institute of Health Policy Management and Evaluation, University of Toronto, Toronto, Ontario, Canada
| | - Rakesh Shiradkar
- Department of Biomedical Engineering, Emory University, Atlanta, Georgia, United States of America
| |
Collapse
|
17
|
Zhang F, Xu Y, Zhou Z, Zhang H, Yang K. Critical element prediction of tracheal intubation difficulty: Automatic Mallampati classification by jointly using handcrafted and attention-based deep features. Comput Biol Med 2022; 150:106182. [PMID: 36242810 DOI: 10.1016/j.compbiomed.2022.106182] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2022] [Revised: 09/14/2022] [Accepted: 10/01/2022] [Indexed: 11/20/2022]
Abstract
Preoperative assessment of the difficulty of tracheal intubation is of great importance in anesthesia practice because failed intubation can lead to severe complications and even death. The Mallampati score is widely used as a critical assessment criterion in combination with other measures to assess the difficulty of tracheal intubation. The performance of existing methods for Mallampati classification with artificial intelligence (AI) is unreliable to the extent that the current clinical judgment of the Mallampati score relies entirely on doctors' experience. In this paper, we propose a new method for automatic Mallampati classification. Our method extracts deep features that are more favorable for the Mallampati classification task by introducing an attention mechanism into the basic deep convolutional neural network (DCNN) and then further improves the classification performance by jointly using attention-based deep features with handcrafted features. We conducted experiments on a dataset consisting of 321 oral images collected online. The proposed method has a classification accuracy of 97.50%, a sensitivity of 96.52%, a specificity of 98.05%, and an F1 score of 96.52% after five-fold cross-validation. The experimental results show that our proposed method is superior to other methods, can assist doctors in determining Mallampati class objectively and accurately, and provide an essential reference element for assessing the difficulty of tracheal intubation.
Collapse
Affiliation(s)
- Fan Zhang
- Unmanned System Research Institute, Northwestern Polytechnical University, Xi'an, 710072, Shaanxi, China.
| | - Yuelei Xu
- Unmanned System Research Institute, Northwestern Polytechnical University, Xi'an, 710072, Shaanxi, China.
| | - Zhaoyun Zhou
- Department of Anesthesiology, Tai'an Central Hospital, Tai'an, 271000, Shandong, China.
| | - Han Zhang
- Department of Anesthesiology, Honghui Hospital, Xi'an Jiaotong University, Xi'an, 710054, Shaanxi, China.
| | - Ke Yang
- Department of Anesthesiology, Fuwai Yunnan Cardiovascular Hospital, Kunming, 650102, Yunnan, China.
| |
Collapse
|
18
|
Kong HJ, Kim JY, Moon HM, Park HC, Kim JW, Lim R, Woo J, Fakhri GE, Kim DW, Kim S. Automation of generative adversarial network-based synthetic data-augmentation for maximizing the diagnostic performance with paranasal imaging. Sci Rep 2022; 12:18118. [PMID: 36302815 PMCID: PMC9613909 DOI: 10.1038/s41598-022-22222-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2022] [Accepted: 10/11/2022] [Indexed: 12/30/2022] Open
Abstract
Thus far, there have been no reported specific rules for systematically determining the appropriate augmented sample size to optimize model performance when conducting data augmentation. In this paper, we report on the feasibility of synthetic data augmentation using generative adversarial networks (GAN) by proposing an automation pipeline to find the optimal multiple of data augmentation to achieve the best deep learning-based diagnostic performance in a limited dataset. We used Waters' view radiographs for patients diagnosed with chronic sinusitis to demonstrate the method developed herein. We demonstrate that our approach produces significantly better diagnostic performance parameters than models trained using conventional data augmentation. The deep learning method proposed in this study could be implemented to assist radiologists in improving their diagnosis. Researchers and industry workers could overcome the lack of training data by employing our proposed automation pipeline approach in GAN-based synthetic data augmentation. This is anticipated to provide new means to overcome the shortage of graphic data for algorithm training.
Collapse
Affiliation(s)
- Hyoun-Joong Kong
- grid.412484.f0000 0001 0302 820XTransdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Jongno-Gu, Seoul, 03080 Republic of Korea ,grid.31501.360000 0004 0470 5905Medical Big Data Research Center, Seoul National University College of Medicine, Jongno-Gu, Seoul, 03080 Republic of Korea ,grid.31501.360000 0004 0470 5905Department of Biomedical Engineering, Seoul National University College of Medicine, 101 Daehak-Ro, Jongno-Gu, Seoul, 03080 Republic of Korea
| | - Jin Youp Kim
- Department of Otorhinolaryngology-Head and Neck Surgery, Ilsan Hospital, Dongguk University, Gyeonggi, 10326 Republic of Korea ,grid.31501.360000 0004 0470 5905Interdisciplinary Program of Medical Informatics, Seoul National University College of Medicine, Seoul, 03080 Republic of Korea
| | - Hye-Min Moon
- grid.31501.360000 0004 0470 5905Interdisciplinary for Bioengineering, Seoul National University, Jongno-Gu, Seoul, 03080 Republic of Korea
| | - Hae Chan Park
- grid.412480.b0000 0004 0647 3378Department of Otorhinolaryngology-Head and Neck Surgery, Seoul National University Bundang Hospital, Gyeonggi, 13620 Republic of Korea
| | - Jeong-Whun Kim
- grid.412480.b0000 0004 0647 3378Department of Otorhinolaryngology-Head and Neck Surgery, Seoul National University Bundang Hospital, Gyeonggi, 13620 Republic of Korea
| | - Ruth Lim
- grid.38142.3c000000041936754XDepartment of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114 USA
| | - Jonghye Woo
- grid.38142.3c000000041936754XDepartment of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114 USA
| | - Georges El Fakhri
- grid.38142.3c000000041936754XDepartment of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA 02114 USA
| | - Dae Woo Kim
- grid.484628.4 0000 0001 0943 2764Department of Otorhinolaryngology-Head and Neck Surgery, Boramae Medical Center, Seoul Metropolitan Government-Seoul National University 20, Boramae-Ro 5-Gil, Dongjak-Gu, Seoul, 07061 Republic of Korea
| | - Sungwan Kim
- grid.412484.f0000 0001 0302 820XTransdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Jongno-Gu, Seoul, 03080 Republic of Korea ,grid.31501.360000 0004 0470 5905Department of Biomedical Engineering, Seoul National University College of Medicine, 101 Daehak-Ro, Jongno-Gu, Seoul, 03080 Republic of Korea ,grid.412484.f0000 0001 0302 820XDepartment of Biomedical Engineering, Seoul National University Hospital, Jongno-Gu, Seoul, 03080 Republic of Korea
| |
Collapse
|
19
|
Huang X, Li Z, Zhang M, Gao S. Fusing hand-crafted and deep-learning features in a convolutional neural network model to identify prostate cancer in pathology images. Front Oncol 2022; 12:994950. [PMID: 36237311 PMCID: PMC9552083 DOI: 10.3389/fonc.2022.994950] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Accepted: 09/09/2022] [Indexed: 11/13/2022] Open
Abstract
Prostate cancer can be diagnosed by prostate biopsy using transectal ultrasound guidance. The high number of pathology images from biopsy tissues is a burden on pathologists, and analysis is subjective and susceptible to inter-rater variability. The use of machine learning techniques could make prostate histopathology diagnostics more precise, consistent, and efficient overall. This paper presents a new classification fusion network model that was created by fusing eight advanced image features: seven hand-crafted features and one deep-learning feature. These features are the scale-invariant feature transform (SIFT), speeded up robust feature (SURF), oriented features from accelerated segment test (FAST) and rotated binary robust independent elementary features (BRIEF) (ORB) of local features, shape and texture features of the cell nuclei, the histogram of oriented gradients (HOG) feature of the cavities, a color feature, and a convolution deep-learning feature. Matching, integrated, and fusion networks are the three essential components of the proposed deep-learning network. The integrated network consists of both a backbone and an additional network. When classifying 1100 prostate pathology images using this fusion network with different backbones (ResNet-18/50, VGG-11/16, and DenseNet-121/201), we discovered that the proposed model with the ResNet-18 backbone achieved the best performance in terms of the accuracy (95.54%), specificity (93.64%), and sensitivity (97.27%) as well as the area under the receiver operating characteristic curve (98.34%). However, each of the assessment criteria for these separate features had a value lower than 90%, which demonstrates that the suggested model combines differently derived characteristics in an effective manner. Moreover, a Grad-CAM++ heatmap was used to observe the differences between the proposed model and ResNet-18 in terms of the regions of interest. This map showed that the proposed model was better at focusing on cancerous cells than ResNet-18. Hence, the proposed classification fusion network, which combines hand-crafted features and a deep-learning feature, is useful for computer-aided diagnoses based on pathology images of prostate cancer. Because of the similarities in the feature engineering and deep learning for different types of pathology images, the proposed method could be used for other pathology images, such as those of breast, thyroid cancer.
Collapse
Affiliation(s)
- Xinrui Huang
- Department of Biochemistry and Biophysics, School of Basic Medical Sciences, Peking University, Beijing, China
| | - Zhaotong Li
- Institute of Medical Technology, Health Science Center, Peking University, Beijing, China
- *Correspondence: Zhaotong Li, ; Song Gao,
| | - Minghui Zhang
- Department of Pathology, Guangdong Provincial People’s Hospital, Guangzhou, China
| | - Song Gao
- Institute of Medical Technology, Health Science Center, Peking University, Beijing, China
- *Correspondence: Zhaotong Li, ; Song Gao,
| |
Collapse
|
20
|
Anterior Cruciate Ligament Tear Detection Based on Deep Convolutional Neural Network. Diagnostics (Basel) 2022; 12:diagnostics12102314. [PMID: 36292003 PMCID: PMC9600338 DOI: 10.3390/diagnostics12102314] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 09/02/2022] [Accepted: 09/07/2022] [Indexed: 11/30/2022] Open
Abstract
Anterior cruciate ligament (ACL) tear is very common in football players, volleyball players, sprinters, runners, etc. It occurs frequently due to extra stretching and sudden movement and causes extreme pain to the patient. Various computer vision-based techniques have been employed for ACL tear detection, but the performance of most of these systems is challenging because of the complex structure of knee ligaments. This paper presents a three-layered compact parallel deep convolutional neural network (CPDCNN) to enhance the feature distinctiveness of the knee MRI images for anterior cruciate ligament (ACL) tear detection in knee MRI images. The performance of the proposed approach is evaluated for the MRNet knee images dataset using accuracy, recall, precision, and the F1 score. The proposed CPDCNN offers an overall accuracy of 96.60%, a recall rate of 0.9668, a precision of 0.9654, and an F1 score of 0.9582, which shows superiority over the existing state-of-the-art methods for knee tear detection.
Collapse
|
21
|
Rashed BM, Popescu N. Critical Analysis of the Current Medical Image-Based Processing Techniques for Automatic Disease Evaluation: Systematic Literature Review. SENSORS (BASEL, SWITZERLAND) 2022; 22:7065. [PMID: 36146414 PMCID: PMC9501859 DOI: 10.3390/s22187065] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/20/2022] [Revised: 09/06/2022] [Accepted: 09/14/2022] [Indexed: 06/16/2023]
Abstract
Medical image processing and analysis techniques play a significant role in diagnosing diseases. Thus, during the last decade, several noteworthy improvements in medical diagnostics have been made based on medical image processing techniques. In this article, we reviewed articles published in the most important journals and conferences that used or proposed medical image analysis techniques to diagnose diseases. Starting from four scientific databases, we applied the PRISMA technique to efficiently process and refine articles until we obtained forty research articles published in the last five years (2017-2021) aimed at answering our research questions. The medical image processing and analysis approaches were identified, examined, and discussed, including preprocessing, segmentation, feature extraction, classification, evaluation metrics, and diagnosis techniques. This article also sheds light on machine learning and deep learning approaches. We also focused on the most important medical image processing techniques used in these articles to establish the best methodologies for future approaches, discussing the most efficient ones and proposing in this way a comprehensive reference source of methods of medical image processing and analysis that can be very useful in future medical diagnosis systems.
Collapse
Affiliation(s)
| | - Nirvana Popescu
- Computer Science Department, University Politehnica of Bucharest, 060042 Bucharest, Romania
| |
Collapse
|
22
|
Zhu J, Geng J, Shan W, Zhang B, Shen H, Dong X, Liu M, Li X, Cheng L. Development and validation of a deep learning model for breast lesion segmentation and characterization in multiparametric MRI. Front Oncol 2022; 12:946580. [PMID: 36033449 PMCID: PMC9402900 DOI: 10.3389/fonc.2022.946580] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2022] [Accepted: 07/12/2022] [Indexed: 11/13/2022] Open
Abstract
Importance The utilization of artificial intelligence for the differentiation of benign and malignant breast lesions in multiparametric MRI (mpMRI) assists radiologists to improve diagnostic performance. Objectives To develop an automated deep learning model for breast lesion segmentation and characterization and to evaluate the characterization performance of AI models and radiologists. Materials and methods For lesion segmentation, 2,823 patients were used for the training, validation, and testing of the VNet-based segmentation models, and the average Dice similarity coefficient (DSC) between the manual segmentation by radiologists and the mask generated by VNet was calculated. For lesion characterization, 3,303 female patients with 3,607 pathologically confirmed lesions (2,213 malignant and 1,394 benign lesions) were used for the three ResNet-based characterization models (two single-input and one multi-input models). Histopathology was used as the diagnostic criterion standard to assess the characterization performance of the AI models and the BI-RADS categorized by the radiologists, in terms of sensitivity, specificity, accuracy, and the area under the receiver operating characteristic curve (AUC). An additional 123 patients with 136 lesions (81 malignant and 55 benign lesions) from another institution were available for external testing. Results Of the 5,811 patients included in the study, the mean age was 46.14 (range 11–89) years. In the segmentation task, a DSC of 0.860 was obtained between the VNet-generated mask and manual segmentation by radiologists. In the characterization task, the AUCs of the multi-input and the other two single-input models were 0.927, 0.821, and 0.795, respectively. Compared to the single-input DWI or DCE model, the multi-input DCE and DWI model obtained a significant increase in sensitivity, specificity, and accuracy (0.831 vs. 0.772/0.776, 0.874 vs. 0.630/0.709, 0.846 vs. 0.721/0.752). Furthermore, the specificity of the multi-input model was higher than that of the radiologists, whether using BI-RADS category 3 or 4 as a cutoff point (0.874 vs. 0.404/0.841), and the accuracy was intermediate between the two assessment methods (0.846 vs. 0.773/0.882). For the external testing, the performance of the three models remained robust with AUCs of 0.812, 0.831, and 0.885, respectively. Conclusions Combining DCE with DWI was superior to applying a single sequence for breast lesion characterization. The deep learning computer-aided diagnosis (CADx) model we developed significantly improved specificity and achieved comparable accuracy to the radiologists with promise for clinical application to provide preliminary diagnoses.
Collapse
Affiliation(s)
- Jingjin Zhu
- School of Medicine, Nankai University, Tianjin, China
- Department of General Surgery, Chinese People’s Liberation Army General Hospital, Beijing, China
| | - Jiahui Geng
- Department of Neurology, Beijing Tiantan Hospital, Beijing, China
| | - Wei Shan
- Department of Neurology, Beijing Tiantan Hospital, Capital Medical University, Beijing, China
| | - Boya Zhang
- School of Medicine, Nankai University, Tianjin, China
- Department of General Surgery, Chinese People’s Liberation Army General Hospital, Beijing, China
| | - Huaqing Shen
- Department of Neurology, Beijing Tiantan Hospital, Beijing, China
| | - Xiaohan Dong
- Department of Radiology, Chinese People’s Liberation Army General Hospital, Beijing, China
| | - Mei Liu
- Department of Pathology, Chinese People’s Liberation Army General Hospital, Beijing, China
| | - Xiru Li
- Department of General Surgery, Chinese People’s Liberation Army General Hospital, Beijing, China
- *Correspondence: Liuquan Cheng, ; Xiru Li,
| | - Liuquan Cheng
- Department of Radiology, Chinese People’s Liberation Army General Hospital, Beijing, China
- *Correspondence: Liuquan Cheng, ; Xiru Li,
| |
Collapse
|
23
|
A Deep Learning Method for Early Detection of Diabetic Foot Using Decision Fusion and Thermal Images. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12157524] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Diabetes mellitus (DM) is one of the major diseases that cause death worldwide and lead to complications of diabetic foot ulcers (DFU). Improper and late handling of a diabetic foot patient can result in an amputation of the patient’s foot. Early detection of DFU symptoms can be observed using thermal imaging with a computer-assisted classifier. Previous study of DFU detection using thermal image only achieved 97% of accuracy, and it has to be improved. This article proposes a novel framework for DFU classification based on thermal imaging using deep neural networks and decision fusion. Here, decision fusion combines the classification result from a parallel classifier. We used the convolutional neural network (CNN) model of ShuffleNet and MobileNetV2 as the baseline classifier. In developing the classifier model, firstly, the MobileNetV2 and ShuffleNet were trained using plantar thermogram datasets. Then, the classification results of those two models were fused using a novel decision fusion method to increase the accuracy rate. The proposed framework achieved 100% accuracy in classifying the DFU thermal images in binary classes of positive and negative cases. The accuracy of the proposed Decision Fusion (DF) was increased by about 3.4% from baseline ShuffleNet and MobileNetV2. Overall, the proposed framework outperformed in classifying the images compared with the state-of-the-art deep learning and the traditional machine-learning-based classifier.
Collapse
|
24
|
Abbas Q. A hybrid transfer learning-based architecture for recognition of medical imaging modalities for healthcare experts. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-212171] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Due to the wide range of diseases and imaging modalities, a retrieving system is a challenging task to access the corresponding clinical cases from a large medical repository on time. Several computer-aided systems (CADx) are developed to recognize medical imaging modalities (MIM) based on various standard machine learning (SML) and advanced deep learning (DL) algorithms. Pre-trained models like convolutional neural networks (CNN) are used in the past as a transfer learning (TL) architecture. However, it is a challenging task to use these pre-trained models for some unseen datasets with a different domain of features. To classify different medical images, the relevant features with a robust classifier are needed and still, it is unsolved task due to MIM-based features. In this paper, a hybrid MIM-based classification system is developed by integrating the pre-trained VGG-19 and ResNet34 models into the original CNN model. Next, the MIM-DTL model is fine-tuned by updating the weights of new layers as well as weights of original CNN layers. The performance of MIM-DTL is compared with state-of-the-art systems based on cancer imaging archive (TCIA), Kvasir and lower extremity radiographs (LERA) datasets in terms of statistical measures such as accuracy (ACC), sensitivity (SE) and specificity (SP). On average, the MIM-DTL model achieved 99% of ACC, SE of 97.5% and SP of 98% along with smaller epochs compare to other TL. The experimental results show that the MIM-DTL model is outperformed to recognize medical imaging modalities and helps the healthcare experts to identify relevant diseases.
Collapse
Affiliation(s)
- Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| |
Collapse
|
25
|
Intelligent Computer-Aided Model for Efficient Diagnosis of Digital Breast Tomosynthesis 3D Imaging Using Deep Learning. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12115736] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Digital breast tomosynthesis (DBT) is a highly promising 3D imaging modality for breast diagnosis. Tissue overlapping is a challenge with traditional 2D mammograms; however, since digital breast tomosynthesis can obtain three-dimensional images, tissue overlapping is reduced, making it easier for radiologists to detect abnormalities and resulting in improved and more accurate diagnosis. In this study, a new computer-aided multi-class diagnosis system is proposed that integrates DBT augmentation and colour feature map with a modified deep learning architecture (Mod_AlexNet). To the proposed modified deep learning architecture (Mod AlexNet), an optimization layer with multiple high performing optimizers is incorporated so that it can be evaluated and optimised using various optimization techniques. Two experimental scenarios are applied, the first scenario proposed a computer-aided diagnosis (CAD) model that integrated DBT augmentation, image enhancement techniques and colour feature mapping with six deep learning models for feature extraction, including ResNet-18, AlexNet, GoogleNet, MobileNetV2, VGG-16 and DenseNet-201, to efficiently classify DBT slices. The second scenario compared the performance of the newly proposed Mod_AlexNet architecture and traditional AlexNet, using several optimization techniques and different evaluation performance metrics were computed. The optimization techniques included adaptive moment estimation (Adam), root mean squared propagation (RMSProp), and stochastic gradient descent with momentum (SGDM), for different batch sizes, including 32, 64 and 512. Experiments have been conducted on a large benchmark dataset of breast tomography scans. The performance of the first scenario was compared in terms of accuracy, precision, sensitivity, specificity, runtime, and f1-score. While in the second scenario, performance was compared in terms of training accuracy, training loss, and test accuracy. In the first scenario, results demonstrated that AlexNet reported improvement rates of 1.69%, 5.13%, 6.13%, 4.79% and 1.6%, compared to ResNet-18, MobileNetV2, GoogleNet, DenseNet-201 and VGG16, respectively. Experimental analysis with different optimization techniques and batch sizes demonstrated that the proposed Mod_AlexNet architecture outperformed AlexNet in terms of test accuracy with improvement rates of 3.23%, 1.79% and 1.34% when compared using SGDM, Adam, and RMSProp optimizers, respectively.
Collapse
|
26
|
A Novel -Gram-Based Image Classification Model and Its Applications in Diagnosing Thyroid Nodule and Retinal OCT Images. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2022; 2022:3151554. [PMID: 35547561 PMCID: PMC9085325 DOI: 10.1155/2022/3151554] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/06/2021] [Revised: 04/14/2022] [Accepted: 04/16/2022] [Indexed: 11/18/2022]
Abstract
Imbalanced classes and dimensional disasters are critical challenges in medical image classification. As a classical machine learning model, the n-gram model has shown excellent performance in addressing this issue in text classification. In this study, we proposed an algorithm to classify medical images by extracting their n-gram semantic features. This algorithm first converts an image classification problem to a text classification problem by building an n-gram corpus for an image. After that, the algorithm was based on the n-gram model to classify images. The algorithm was evaluated by two independent public datasets. The first experiment is to diagnose benign and malignant thyroid nodules. The best area under the curve (AUC) is 0.989. The second experiment is to diagnose the type of fundus lesion. The best result is that it correctly identified 86.667% of patients with dry age-related macular degeneration (AMD), 93.333% of patients with diabetic macular edema (DME), and 93.333% of normal individuals.
Collapse
|
27
|
Fok WYR, Grashei M, Skinner JG, Menze BH, Schilling F. Prediction of multiple pH compartments by deep learning in magnetic resonance spectroscopy with hyperpolarized 13C-labelled zymonic acid. EJNMMI Res 2022; 12:24. [PMID: 35460436 PMCID: PMC9035201 DOI: 10.1186/s13550-022-00894-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Accepted: 04/05/2022] [Indexed: 11/11/2022] Open
Abstract
BACKGROUND Hyperpolarization enhances the sensitivity of nuclear magnetic resonance experiments by between four and five orders of magnitude. Several hyperpolarized sensor molecules have been introduced that enable high sensitivity detection of metabolism and physiological parameters. However, hyperpolarized magnetic resonance spectroscopy imaging (MRSI) often suffers from poor signal-to-noise ratio and spectral analysis is complicated by peak overlap. Here, we study measurements of extracellular pH (pHe) by hyperpolarized zymonic acid, where multiple pHe compartments, such as those observed in healthy kidney or other heterogeneous tissue, result in a cluster of spectrally overlapping peaks, which is hard to resolve with conventional spectroscopy analysis routines. METHODS We investigate whether deep learning methods can yield improved pHe prediction in hyperpolarized zymonic acid spectra of multiple pHe compartments compared to conventional line fitting. As hyperpolarized 13C-MRSI data sets are often small, a convolutional neural network (CNN) and a multilayer perceptron (MLP) were trained with either a synthetic or a mixed (synthetic and augmented) data set of acquisitions from the kidneys of healthy mice. RESULTS Comparing the networks' performances compartment-wise on a synthetic test data set and eight real kidney data shows superior performance of CNN compared to MLP and equal or superior performance compared to conventional line fitting. For correct prediction of real kidney pHe values, training with a mixed data set containing only 0.5% real data shows a large improvement compared to training with synthetic data only. Using a manual segmentation approach, pH maps of kidney compartments can be improved by neural network predictions for voxels including three pH compartments. CONCLUSION The results of this study indicate that CNNs offer a reliable, accurate, fast and non-interactive method for analysis of hyperpolarized 13C MRS and MRSI data, where low amounts of acquired data can be complemented to achieve suitable network training.
Collapse
Affiliation(s)
- Wai-Yan Ryana Fok
- Department of Informatics, Technical University of Munich, 85748, Garching, Germany
| | - Martin Grashei
- Department of Nuclear Medicine, TUM School of Medicine, Klinikum Rechts der Isar, Technical University of Munich, 81675, Munich, Germany
| | - Jason G Skinner
- Department of Nuclear Medicine, TUM School of Medicine, Klinikum Rechts der Isar, Technical University of Munich, 81675, Munich, Germany
| | - Bjoern H Menze
- Department of Informatics, Technical University of Munich, 85748, Garching, Germany
| | - Franz Schilling
- Department of Nuclear Medicine, TUM School of Medicine, Klinikum Rechts der Isar, Technical University of Munich, 81675, Munich, Germany.
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany.
| |
Collapse
|
28
|
Deep learning-enabled mobile application for efficient and robust herb image recognition. Sci Rep 2022; 12:6579. [PMID: 35449192 PMCID: PMC9023495 DOI: 10.1038/s41598-022-10449-9] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2021] [Accepted: 04/06/2022] [Indexed: 12/31/2022] Open
Abstract
With the increasing popularity of herbal medicine, high standards of the high quality control of herbs becomes a necessity, with the herb recognition as one of the great challenges. Due to the complicated processing procedure of the herbs, methods of manual recognition that require chemical materials and expert knowledge, such as fingerprint and experience, have been used. Automatic methods can partially alleviate the problem by deep learning based herb image recognition, but most studies require powerful and expensive computation hardware, which is not friendly to resource-limited settings. In this paper, we introduce a deep learning-enabled mobile application which can run entirely on common low-cost smartphones for efficient and robust herb image recognition with a quite competitive recognition accuracy in resource-limited situations. We hope this application can make contributions to the increasing accessibility of herbal medicine worldwide.
Collapse
|
29
|
FF-PCA-LDA: Intelligent Feature Fusion Based PCA-LDA Classification System for Plant Leaf Diseases. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12073514] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
Crop leaf disease management and control pose significant impact on enhancement in yield and quality to fulfill consumer needs. For smart agriculture, an intelligent leaf disease identification system is inevitable for efficient crop health monitoring. In this view, a novel approach is proposed for crop disease identification using feature fusion and PCA-LDA classification (FF-PCA-LDA). Handcrafted hybrid and deep features are extracted from RGB images. TL-ResNet50 is used to extract the deep features. Fused feature vector is obtained by combining handcrafted hybrid and deep features. After fusing the image features, PCA is employed to select most discriminant features for LDA model development. Potato crop leaf disease identification is used as a case study for the validation of the approach. The developed system is experimentally validated on a potato crop leaf benchmark dataset. It offers high accuracy of 98.20% on an unseen dataset which was not used during the model training process. Performance comparison of the proposed technique with other approaches shows its superiority. Owing to the better discrimination and learning ability, the proposed approach overcomes the leaf segmentation step. The developed approach may be used as an automated tool for crop monitoring, management control, and can be extended for other crop types.
Collapse
|
30
|
Li Z, Wu F, Hong F, Gai X, Cao W, Zhang Z, Yang T, Wang J, Gao S, Peng C. Computer-Aided Diagnosis of Spinal Tuberculosis From CT Images Based on Deep Learning With Multimodal Feature Fusion. Front Microbiol 2022; 13:823324. [PMID: 35283815 PMCID: PMC8905347 DOI: 10.3389/fmicb.2022.823324] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2021] [Accepted: 01/13/2022] [Indexed: 11/13/2022] Open
Abstract
Background Spinal tuberculosis (TB) has the highest incidence in remote plateau areas, particularly in Tibet, China, due to inadequate local healthcare services, which not only facilitates the transmission of TB bacteria but also increases the burden on grassroots hospitals. Computer-aided diagnosis (CAD) is urgently required to improve the efficiency of clinical diagnosis of TB using computed tomography (CT) images. However, classical machine learning with handcrafted features generally has low accuracy, and deep learning with self-extracting features relies heavily on the size of medical datasets. Therefore, CAD, which effectively fuses multimodal features, is an alternative solution for spinal TB detection. Methods A new deep learning method is proposed that fuses four elaborate image features, specifically three handcrafted features and one convolutional neural network (CNN) feature. Spinal TB CT images were collected from 197 patients with spinal TB, from 2013 to 2020, in the People’s Hospital of Tibet Autonomous Region, China; 3,000 effective lumbar spine CT images were randomly screened to our dataset, from which two sets of 1,500 images each were classified as tuberculosis (positive) and health (negative). In addition, virtual data augmentation is proposed to enlarge the handcrafted features of the TB dataset. Essentially, the proposed multimodal feature fusion CNN consists of four main sections: matching network, backbone (ResNet-18/50, VGG-11/16, DenseNet-121/161), fallen network, and gated information fusion network. Detailed performance analyses were conducted based on the multimodal features, proposed augmentation, model stability, and model-focused heatmap. Results Experimental results showed that the proposed model with VGG-11 and virtual data augmentation exhibited optimal performance in terms of accuracy, specificity, sensitivity, and area under curve. In addition, an inverse relationship existed between the model size and test accuracy. The model-focused heatmap also shifted from the irrelevant region to the bone destruction caused by TB. Conclusion The proposed augmentation effectively simulated the real data distribution in the feature space. More importantly, all the evaluation metrics and analyses demonstrated that the proposed deep learning model exhibits efficient feature fusion for multimodal features. Our study provides a profound insight into the preliminary auxiliary diagnosis of spinal TB from CT images applicable to the Tibetan area.
Collapse
Affiliation(s)
- Zhaotong Li
- Institute of Medical Technology, Peking University Health Science Center, Beijing, China.,School of Health Humanities, Peking University, Beijing, China
| | - Fengliang Wu
- Beijing Key Laboratory of Spinal Disease Research, Engineering Research Center of Bone and Joint Precision Medicine, Department of Orthopedics, Peking University Third Hospital, Beijing, China.,Department of Orthopedic, People's Hospital of Tibet Autonomous Region, Lhasa, China
| | - Fengze Hong
- Medical College, Tibet University, Lhasa, China
| | - Xiaoyan Gai
- Department of Respiratory and Critical Care Medicine, Peking University Third Hospital, Beijing, China
| | - Wenli Cao
- Tuberculosis Department, Beijing Geriatric Hospital, Beijing, China
| | - Zeru Zhang
- Institute of Medical Technology, Peking University Health Science Center, Beijing, China.,School of Health Humanities, Peking University, Beijing, China
| | - Timin Yang
- Department of Orthopedic, People's Hospital of Tibet Autonomous Region, Lhasa, China
| | - Jiu Wang
- Department of Orthopedic, People's Hospital of Tibet Autonomous Region, Lhasa, China
| | - Song Gao
- Institute of Medical Technology, Peking University Health Science Center, Beijing, China
| | - Chao Peng
- Department of Orthopedic, People's Hospital of Tibet Autonomous Region, Lhasa, China
| |
Collapse
|
31
|
Montalbo FJ. Truncating fined-tuned vision-based models to lightweight deployable diagnostic tools for SARS-CoV-2 infected chest X-rays and CT-scans. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 81:16411-16439. [PMID: 35261555 PMCID: PMC8893243 DOI: 10.1007/s11042-022-12484-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/08/2021] [Revised: 10/05/2021] [Accepted: 01/25/2022] [Indexed: 06/14/2023]
Abstract
In such a brief period, the recent coronavirus (COVID-19) already infected large populations worldwide. Diagnosing an infected individual requires a Real-Time Polymerase Chain Reaction (RT-PCR) test, which can become expensive and limited in most developing countries, making them rely on alternatives like Chest X-Rays (CXR) or Computerized Tomography (CT) scans. However, results from these imaging approaches radiated confusion for medical experts due to their similarities with other diseases like pneumonia. Other solutions based on Deep Convolutional Neural Network (DCNN) recently improved and automated the diagnosis of COVID-19 from CXRs and CT scans. However, upon examination, most proposed studies focused primarily on accuracy rather than deployment and reproduction, which may cause them to become difficult to reproduce and implement in locations with inadequate computing resources. Therefore, instead of focusing only on accuracy, this work investigated the effects of parameter reduction through a proposed truncation method and analyzed its effects. Various DCNNs had their architectures truncated, which retained only their initial core block, reducing their parameter sizes to <1 M. Once trained and validated, findings have shown that a DCNN with robust layer aggregations like the InceptionResNetV2 had less vulnerability to the adverse effects of the proposed truncation. The results also showed that from its full-length size of 55 M with 98.67% accuracy, the proposed truncation reduced its parameters to only 441 K and still attained an accuracy of 97.41%, outperforming other studies based on its size to performance ratio.
Collapse
Affiliation(s)
- Francis Jesmar Montalbo
- College of Informatics and Computing Sciences, Batangas State University, Rizal Avenue Extension, Batangas, Batangas City, Philippines
| |
Collapse
|
32
|
Yousef R, Gupta G, Yousef N, Khari M. A holistic overview of deep learning approach in medical imaging. MULTIMEDIA SYSTEMS 2022; 28:881-914. [PMID: 35079207 PMCID: PMC8776556 DOI: 10.1007/s00530-021-00884-5] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2021] [Accepted: 12/23/2021] [Indexed: 05/07/2023]
Abstract
Medical images are a rich source of invaluable necessary information used by clinicians. Recent technologies have introduced many advancements for exploiting the most of this information and use it to generate better analysis. Deep learning (DL) techniques have been empowered in medical images analysis using computer-assisted imaging contexts and presenting a lot of solutions and improvements while analyzing these images by radiologists and other specialists. In this paper, we present a survey of DL techniques used for variety of tasks along with the different medical image's modalities to provide critical review of the recent developments in this direction. We have organized our paper to provide significant contribution of deep leaning traits and learn its concepts, which is in turn helpful for non-expert in medical society. Then, we present several applications of deep learning (e.g., segmentation, classification, detection, etc.) which are commonly used for clinical purposes for different anatomical site, and we also present the main key terms for DL attributes like basic architecture, data augmentation, transfer learning, and feature selection methods. Medical images as inputs to deep learning architectures will be the mainstream in the coming years, and novel DL techniques are predicted to be the core of medical images analysis. We conclude our paper by addressing some research challenges and the suggested solutions for them found in literature, and also future promises and directions for further developments.
Collapse
Affiliation(s)
- Rammah Yousef
- Yogananda School of AI Computer and Data Sciences, Shoolini University, Solan, 173229 Himachal Pradesh India
| | - Gaurav Gupta
- Yogananda School of AI Computer and Data Sciences, Shoolini University, Solan, 173229 Himachal Pradesh India
| | - Nabhan Yousef
- Electronics and Communication Engineering, Marwadi University, Rajkot, Gujrat India
| | - Manju Khari
- Jawaharlal Nehru University, New Delhi, India
| |
Collapse
|
33
|
Li J, Wang P, Zhou Y, Liang H, Lu Y, Luan K. A novel classification method of lymph node metastasis in colorectal cancer. Bioengineered 2021; 12:2007-2021. [PMID: 34024255 PMCID: PMC8806456 DOI: 10.1080/21655979.2021.1930333] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2021] [Revised: 05/07/2021] [Accepted: 05/08/2021] [Indexed: 11/21/2022] Open
Abstract
Colorectal cancer lymph node metastasis, which is highly associated with the patient's cancer recurrence and survival rate, has been the focus of many therapeutic strategies that are highly associated with the patient's cancer recurrence and survival rate. The popular methods for classification of lymph node metastasis by neural networks, however, show limitations as the available low-level features are inadequate for classification, and the radiologists are unable to quickly review the images. Identifying lymph node metastasis in colorectal cancer is a key factor in the treatment of patients with colorectal cancer. In the present work, an automatic classification method based on deep transfer learning was proposed. Specifically, the method resolved the problem of repetition of low-level features and combined these features with high-level features into a new feature map for classification; and a merged layer which merges all transmitted features from previous layers into a map of the first full connection layer. With a dataset collected from Harbin Medical University Cancer Hospital, the experiment involved a sample of 3,364 patients. Among these samples, 1,646 were positive, and 1,718 were negative. The experiment results showed the sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) were 0.8732, 0.8746, 0.8746 and 0.8728, respectively, and the accuracy and AUC were 0.8358 and 0.8569, respectively. These demonstrated that our method significantly outperformed the previous classification methods for colorectal cancer lymph node metastasis without increasing the depth and width of the model.
Collapse
Affiliation(s)
- Jin Li
- College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin, Heilongjiang Province, China
| | - Peng Wang
- College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin, Heilongjiang Province, China
| | - Yang Zhou
- College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin, Heilongjiang Province, China
- Department of Radiology, Harbin Medical University Cancer Hospital, Harbin, Heilongjiang Province, China
| | - Hong Liang
- College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin, Heilongjiang Province, China
| | - Yang Lu
- College of Information and Electrical Engineering, Heilongjiang Bayi Agricultural University, Daqing, Heilongjiang Province, China
| | - Kuan Luan
- College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin, Heilongjiang Province, China
| |
Collapse
|
34
|
Shakarami A, Menhaj MB, Tarrah H. Diagnosing COVID-19 disease using an efficient CAD system. OPTIK 2021; 241:167199. [PMID: 34028466 PMCID: PMC8130607 DOI: 10.1016/j.ijleo.2021.167199] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2021] [Accepted: 05/12/2021] [Indexed: 05/28/2023]
Abstract
Todays, COVID-19 has caused much death and its spreading speed is increasing, regarding virus mutation. This outbreak warns diagnosing infected people is an important issue. So, in this research, a computer-aided diagnosis (CAD) system called COV-CAD is proposed for diagnosing COVID-19 disease. This COV-CAD system is created by a feature extractor, a classification method, and a content-based imaged retrieval (CBIR) system. The proposed feature extractor is created by using the modified AlexNet CNN. The first modification changes ReLU activation functions to LeakyReLU for increasing efficiency. The second change is converting a fully connected (FC) layer of AlexNet CNN with a new FC, which results in reducing learnable parameters and training time. Another FC layer with dimensions 1 × 64 is added at the end of the feature extractor as the feature vector. In the classification section, a new classification method is defined in which the majority voting technique is applied on outputs of CBIR, SVM, KNN, and Random Forest for final diagnosing. Furthermore, in retrieval section, the proposed method uses CBIR because of its ability to retrieve the most similar images to the image of a patient. Since this feature helps physicians to find the most similar cases, they could conduct further statistical evaluations on profiles of similar patients. The system has been evaluated by accuracy, sensitivity, specificity, F1-score, and mean average precision and its accuracy for CT and X-ray datasets is 93.20% and 99.38%, respectively. The results demonstrate that the proposed method is more efficient than other similar studies.
Collapse
Affiliation(s)
- Ashkan Shakarami
- Department of Computer Engineering, Afarinesh Institute of Higher Education, Borujerd, Iran
| | - Mohammad Bagher Menhaj
- Department of Electrical Engineering, Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran
| | - Hadis Tarrah
- Department of Electrical, Computer and Biomedical Engineering, Islamic Azad University, Qazvin, Iran
| |
Collapse
|
35
|
Duwairi RM, Al-Zboon SA, Al-Dwairi RA, Obaidi A. A Deep Learning Model and a Dataset for Diagnosing Ophthalmology Diseases. JOURNAL OF INFORMATION & KNOWLEDGE MANAGEMENT 2021. [DOI: 10.1142/s0219649221500362] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
The rapid development of artificial neural network techniques, especially convolutional neural networks, encouraged the researchers to adapt such techniques in the medical domain. Specifically, to provide assist tools to help the professionals in patients’ diagnosis. The main problem faced by the researchers in the medical domain is the lack of available annotated datasets which can be used to train and evaluate large and complex deep neural networks. In this paper, to assist researchers who are interested in applying deep learning techniques to aid the ophthalmologists in diagnosing eye-related diseases, we provide an optical coherence tomography dataset with collaboration with ophthalmologists from the King Abdullah University Hospital, Irbid, Jordan. This dataset consists of 21,991 OCT images distributed over seven eye diseases in addition to normal images (no disease), namely, Choroidal Neovascularisation, Full Macular Hole (Full Thickness), Partial Macular Hole, Central Serous Retinopathy, Geographic atrophy, Macular Retinal Oedema, and Vitreomacular Traction. To the best of our knowledge, this dataset is the largest of its kind, where images belong to actual patients from Jordan and the annotation was carried out by ophthalmologists. Two classification tasks were applied to this dataset; a binary classification to distinguish between images which belong to healthy eyes (normal) and images which belong to diseased eyes (abnormal). The second classification task is a multi-class classification, where the deep neural network is trained to distinguish between the seven diseases listed above in addition to the normal case. In both classification tasks, the U-Net neural network was modified and subsequently utilised. This modification adds an additional block of layers to the original U-Net model to become capable of handling classification as the original network is used for image segmentation. The results of the binary classification were equal to 84.90% and 69.50% as accuracy and quadratic weighted kappa, respectively. The results of the multi-class classification, by contrast, were equal to 63.68% and 66.06% as accuracy and quadratic weighted kappa, respectively.
Collapse
Affiliation(s)
- Rehab M. Duwairi
- Department of Computer Information Systems, Jordan University of Science and Technology, Irbid, Jordan
| | - Saad A. Al-Zboon
- Department of Computer Science, Jordan University of Science and Technology, Irbid, Jordan
| | - Rami A. Al-Dwairi
- Division of Ophthalmology, Department of Special Surgery, Faculty of Medicine, Jordan University of Science and Technology, Irbid, Jordan
| | - Ahmad Obaidi
- King Abdullah University Hospital, Irbid, Jordan
| |
Collapse
|
36
|
Sun MD, Halpern AC. Advances in the Etiology, Detection, and Clinical Management of Seborrheic Keratoses. Dermatology 2021; 238:205-217. [PMID: 34311463 DOI: 10.1159/000517070] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Accepted: 05/06/2021] [Indexed: 11/19/2022] Open
Abstract
Seborrheic keratoses (SKs) are ubiquitous, generally benign skin tumors that exhibit high clinical variability. While age is a known risk factor, the precise roles of UV exposure and immune abnormalities are currently unclear. The underlying mechanisms of this benign disorder are paradoxically driven by oncogenic mutations and may have profound implications for our understanding of the malignant state. Advances in molecular pathogenesis suggest that inhibition of Akt and APP, as well as existing treatments for skin cancer, may have therapeutic potential in SK. Dermoscopic criteria have also become increasingly important to the accurate detection of SK, and other noninvasive diagnostic methods, such as reflectance confocal microscopy and optical coherence tomography, are rapidly developing. Given their ability to mimic malignant tumors, SK cases are often used to train artificial intelligence-based algorithms in the computerized detection of skin disease. These technologies are becoming increasingly accurate and have the potential to significantly augment clinical practice. Current treatment options for SK cause discomfort and can lead to adverse post-treatment effects, especially in skin of color. In light of the discontinuation of ESKATA in late 2019, promising alternatives, such as nitric-zinc and trichloroacetic acid topicals, should be further developed. There is also a need for larger, head-to-head trials of emerging laser therapies to ensure that future treatment standards address diverse patient needs.
Collapse
Affiliation(s)
- Mary D Sun
- Icahn School of Medicine at Mount Sinai, New York, New York, USA,
| | - Allan C Halpern
- Dermatology Service, Memorial Sloan Kettering, New York, New York, USA
| |
Collapse
|
37
|
Altaf F, Islam SMS, Janjua NK. A novel augmented deep transfer learning for classification of COVID-19 and other thoracic diseases from X-rays. Neural Comput Appl 2021; 33:14037-14048. [PMID: 33948047 PMCID: PMC8083924 DOI: 10.1007/s00521-021-06044-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2020] [Accepted: 04/13/2021] [Indexed: 12/24/2022]
Abstract
Deep learning has provided numerous breakthroughs in natural imaging tasks. However, its successful application to medical images is severely handicapped with the limited amount of annotated training data. Transfer learning is commonly adopted for the medical imaging tasks. However, a large covariant shift between the source domain of natural images and target domain of medical images results in poor transfer learning. Moreover, scarcity of annotated data for the medical imaging tasks causes further problems for effective transfer learning. To address these problems, we develop an augmented ensemble transfer learning technique that leads to significant performance gain over the conventional transfer learning. Our technique uses an ensemble of deep learning models, where the architecture of each network is modified with extra layers to account for dimensionality change between the images of source and target data domains. Moreover, the model is hierarchically tuned to the target domain with augmented training data. Along with the network ensemble, we also utilize an ensemble of dictionaries that are based on features extracted from the augmented models. The dictionary ensemble provides an additional performance boost to our method. We first establish the effectiveness of our technique with the challenging ChestXray-14 radiography data set. Our experimental results show more than 50% reduction in the error rate with our method as compared to the baseline transfer learning technique. We then apply our technique to a recent COVID-19 data set for binary and multi-class classification tasks. Our technique achieves 99.49% accuracy for the binary classification, and 99.24% for multi-class classification.
Collapse
Affiliation(s)
- Fouzia Altaf
- School of Science, Edith Cowan University, Joondalup, WA Australia
| | - Syed M. S. Islam
- School of Science, Edith Cowan University, Joondalup, WA Australia
| | | |
Collapse
|
38
|
Prewitt Logistic Deep Recurrent Neural Learning for Face Log Detection by Extracting Features from Images. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2021. [DOI: 10.1007/s13369-021-05609-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
39
|
Menaka R, Ramesh R, Dhanagopal R. Aggregation of Region-based and Boundary-based Knowledge Biased Segmentation for Osteoporosis Detection from X-Ray, Dual X-Ray and CT Images. Curr Med Imaging 2021; 17:288-295. [PMID: 32748751 DOI: 10.2174/1573405616999200730175526] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2020] [Revised: 06/03/2020] [Accepted: 06/19/2020] [Indexed: 11/22/2022]
Abstract
BACKGROUND Osteoporosis is a term used to represent the reduced bone density, which is caused by insufficient bone tissue production to balance the old bone tissue removal. Medical Imaging procedures such as X-Ray, Dual X-Ray and Computed Tomography (CT) scans are used widely in osteoporosis diagnosis. There are several existing procedures in practice to assist osteoporosis diagnosis, which can operate using a single imaging method. OBJECTIVE The purpose of this proposed work is to introduce a framework to assist the diagnosis of osteoporosis based on consenting all these X-Ray, Dual X-Ray and CT scan imaging techniques. The proposed work named "Aggregation of Region-based and Boundary-based Knowledge biased Segmentation for Osteoporosis Detection from X-Ray, Dual X-Ray and CT images" (ARBKSOD) is the integration of three functional modules. METHODS Fuzzy Histogram Medical Image Classifier (FHMIC), Log-Gabor Transform based ANN Training for osteoporosis detection (LGTAT) and Knowledge biased Osteoporosis Analyzer (KOA). RESULTS Together, all these three modules make the proposed method ARBKSOD scored the maximum accuracy of 93.11%, the highest precision value of 93.91% while processing the 6th image batch, the highest sensitivity of 92.93%, the highest specificity of 93.79% is observed during the experiment by ARBKSOD while processing the 6th image batch. The best average processing time of 10244 mS is achieved by ARBKSOD while processing the 7th image batch. CONCLUSION Together, all these three modules make the proposed method ARBKSOD to produce a better result.
Collapse
Affiliation(s)
- R Menaka
- Department of Chennai Institute of Technology, Chennai, Tamil Nadu, India
| | - R Ramesh
- Department of Chennai Institute of Technology, Chennai, Tamil Nadu, India
| | - R Dhanagopal
- Department of Chennai Institute of Technology, Chennai, Tamil Nadu, India
| |
Collapse
|
40
|
Chugh G, Kumar S, Singh N. Survey on Machine Learning and Deep Learning Applications in Breast Cancer Diagnosis. Cognit Comput 2021. [DOI: 10.1007/s12559-020-09813-6] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
|
41
|
Multilayer perceptron based deep neural network for early detection of coronary heart disease. HEALTH AND TECHNOLOGY 2020. [DOI: 10.1007/s12553-020-00509-3] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
42
|
Extracting Lungs from CT Images via Deep Convolutional Neural Network Based Segmentation and Two-Pass Contour Refinement. J Digit Imaging 2020; 33:1465-1478. [PMID: 33057882 DOI: 10.1007/s10278-020-00388-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2019] [Revised: 08/17/2020] [Accepted: 09/14/2020] [Indexed: 10/23/2022] Open
Abstract
Lung segmentation is a key step of thoracic computed tomography (CT) image processing, and it plays an important role in computer-aided pulmonary disease diagnostics. However, the presence of image noises, pathologies, vessels, individual anatomical varieties, and so on makes lung segmentation a complex task. In this paper, we present a fully automatic algorithm for segmenting lungs from thoracic CT images accurately. An input image is first spilt into a set of non-overlapping fixed-sized image patches, and a deep convolutional neural network model is constructed to extract initial lung regions by classifying image patches. Superpixel segmentation is then performed on the preprocessed thoracic CT image, and the lung contours are locally refined according to corresponding superpixel contours with our adjacent point statistics method. Segmented lung contours are further globally refined by an edge direction tracing technique for the inclusion of juxta-pleural lesions. Our algorithm is tested on a group of thoracic CT scans with interstitial lung diseases. Experiments show that our algorithm creates an average Dice similarity coefficient of 97.95% and Jaccard's similarity index of 94.48%, with 2.8% average over-segmentation rate and 3.3% under-segmentation rate compared with manually segmented results. Meanwhile, it shows better performance compared with several feature-based machine learning methods and current methods on lung segmentation.
Collapse
|
43
|
MRI Brain Classification Using the Quantum Entropy LBP and Deep-Learning-Based Features. ENTROPY 2020; 22:e22091033. [PMID: 33286802 PMCID: PMC7597092 DOI: 10.3390/e22091033] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/21/2020] [Revised: 09/10/2020] [Accepted: 09/11/2020] [Indexed: 01/10/2023]
Abstract
Brain tumor detection at early stages can increase the chances of the patient’s recovery after treatment. In the last decade, we have noticed a substantial development in the medical imaging technologies, and they are now becoming an integral part in the diagnosis and treatment processes. In this study, we generalize the concept of entropy difference defined in terms of Marsaglia formula (usually used to describe two different figures, statues, etc.) by using the quantum calculus. Then we employ the result to extend the local binary patterns (LBP) to get the quantum entropy LBP (QELBP). The proposed study consists of two approaches of features extractions of MRI brain scans, namely, the QELBP and the deep learning DL features. The classification of MRI brain scan is improved by exploiting the excellent performance of the QELBP–DL feature extraction of the brain in MRI brain scans. The combining all of the extracted features increase the classification accuracy of long short-term memory network when using it as the brain tumor classifier. The maximum accuracy achieved for classifying a dataset comprising 154 MRI brain scan is 98.80%. The experimental results demonstrate that combining the extracted features improves the performance of MRI brain tumor classification.
Collapse
|
44
|
|
45
|
Liu C, Zhao R, Xie W, Pang M. Pathological lung segmentation based on random forest combined with deep model and multi-scale superpixels. Neural Process Lett 2020; 52:1631-1649. [PMID: 32837245 PMCID: PMC7413019 DOI: 10.1007/s11063-020-10330-8] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
Accurate segmentation of lungs in pathological thoracic computed tomography (CT) scans plays an important role in pulmonary disease diagnosis. However, it is still a challenging task due to the variability of pathological lung appearances and shapes. In this paper, we proposed a novel segmentation algorithm based on random forest (RF), deep convolutional network, and multi-scale superpixels for segmenting pathological lungs from thoracic CT images accurately. A pathological thoracic CT image is first segmented based on multi-scale superpixels, and deep features, texture, and intensity features extracted from superpixels are taken as inputs of a group of RF classifiers. With the fusion of classification results of RFs by a fractional-order gray correlation approach, we capture an initial segmentation of pathological lungs. We finally utilize a divide-and-conquer strategy to deal with segmentation refinement combining contour correction of left lungs and region repairing of right lungs. Our algorithm is tested on a group of thoracic CT images affected with interstitial lung diseases. Experiments show that our algorithm can achieve a high segmentation accuracy with an average DSC of 96.45% and PPV of 95.07%. Compared with several existing lung segmentation methods, our algorithm exhibits a robust performance on pathological lung segmentation. Our algorithm can be employed reliably for lung field segmentation of pathologic thoracic CT images with a high accuracy, which is helpful to assist radiologists to detect the presence of pulmonary diseases and quantify its shape and size in regular clinical practices.
Collapse
Affiliation(s)
- Caixia Liu
- Institute of EduInfo Science and Engineering, Nanjing Normal University, Nanjing, China
| | - Ruibin Zhao
- Institute of EduInfo Science and Engineering, Nanjing Normal University, Nanjing, China
| | - Wangli Xie
- Institute of EduInfo Science and Engineering, Nanjing Normal University, Nanjing, China
| | - Mingyong Pang
- Institute of EduInfo Science and Engineering, Nanjing Normal University, Nanjing, China
| |
Collapse
|
46
|
Wady SH, Yousif RZ, Hasan HR. A Novel Intelligent System for Brain Tumor Diagnosis Based on a Composite Neutrosophic-Slantlet Transform Domain for Statistical Texture Feature Extraction. BIOMED RESEARCH INTERNATIONAL 2020; 2020:8125392. [PMID: 32733955 PMCID: PMC7369660 DOI: 10.1155/2020/8125392] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/19/2019] [Revised: 04/10/2020] [Accepted: 06/08/2020] [Indexed: 12/28/2022]
Abstract
Discrete wavelet transform (DWT) is often implemented by an iterative filter bank; hence, a lake of optimization of a discrete time basis is observed with respect to time localization for a constant number of zero moments. This paper discusses and presents an improved form of DWT for feature extraction, called Slantlet transform (SLT) along with neutrosophy, a generalization of fuzzy logic, which is a relatively new logic. Thus, a novel composite NS-SLT model has been suggested as a source to derive statistical texture features that used to identify the malignancy of brain tumor. The MR images in the neutrosophic domain are defined using three membership sets, true (T), false (F), and indeterminate (I); then, SLT was applied to each membership set. Three statistical measurement-based methods are used to extract texture features from images of brain MRI. One-way ANOVA has been applied as a method of reducing the number of extracted features for the classifiers; then, the extracted features are subsequently provided to the four neural network classification techniques, Support Vector Machine Neural Network (SVM-NN), Decision Tree Neural Network (DT-NN), K-Nearest Neighbor Neural Network (KNN-NN), and Naive Bayes Neural Networks (NB-NN), to predict the type of the brain tumor. Meanwhile, the performance of the proposed model is assessed by calculating average accuracy, precision, sensitivity, specificity, and Area Under the Curve (AUC) of the Receiver Operating Characteristic (ROC) curve. The experimental results demonstrate that the proposed approach is quite accurate and efficient for diagnosing brain tumors when the Gray Level Run Length Matrix (GLRLM) features derived from the composite NS-SLT technique is used.
Collapse
Affiliation(s)
- Shakhawan H. Wady
- Applied Computer, College of Medicals and Applied Sciences, Charmo University, Chamchamal, Sulaimani, KRG, Iraq
- Technical College of Informatics, Sulaimani Polytechnic University, Sulaimani, KRG, Iraq
- Department of Information Technology, University College of Goizha, Sulaimani, KRG, Iraq
| | - Raghad Z. Yousif
- Department of Physics, College of Science, Salahaddin University, Erbil, KRG, Iraq
- Department of IT, College of Information Technology, Catholic University in Erbil, KRG, Iraq
| | - Harith R. Hasan
- Department of Computer Science, Kurdistan Technical Institute, Sulaimani, KRG, Iraq
- Computer Science Institute, Sulaimani Polytechnic University, Sulaimani, KRG, Iraq
| |
Collapse
|
47
|
Maroof N, Khan A, Qureshi SA, Rehman AU, Khalil RK, Shim SO. Mitosis detection in breast cancer histopathology images using hybrid feature space. Photodiagnosis Photodyn Ther 2020; 31:101885. [PMID: 32565178 DOI: 10.1016/j.pdpdt.2020.101885] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2020] [Revised: 04/13/2020] [Accepted: 06/12/2020] [Indexed: 12/18/2022]
Abstract
Breast Cancer grading is a challenging task as regards image analysis, which is normally based on mitosis count rate. The mitotic count provides an estimate of aggressiveness of the tumor. The detection of mitosis is a challenging task because in a frame of slides at X40 magnification, there are hundreds of nuclei containing few mitotic nuclei. However, manual counting of mitosis by pathologists is a difficult and time intensive job, moreover conventional method rely mainly on the shape, color, and/or texture features as well as pathologist experience. The objective of this study is to accept the atypaia-2014 mitosis detection challenge, automate the process of mitosis detection and a proposal of a hybrid feature space that provides better discrimination of mitotic and non-mitotic nuclei by combining color features with morphological and texture features. To exploit color channels, they were first selected, and then normalized and cumulative histograms were computed in wavelet domain. A detailed analysis presented on these features in different color channels of respective color spaces using Random Forest (RF) and Support Vector Machine (SVM) classifiers. The proposed hybrid feature space when used with SVM classifier achieved a detection rate of 78.88% and F-measure of 72.07%. Our results, especially high detection rate, indicate that proposed hybrid feature space model contains discriminant information for mitotic nuclei, being therefore a very capable are for exploration to improve the quality of the diagnostic assistance in histopathology.
Collapse
Affiliation(s)
- Noorulain Maroof
- Department of Computer and Information Sciences, Pakistan Institute of Engineering and Applied Sciences, Nilore, 45650, Islamabad, Pakistan
| | - Asifullah Khan
- Department of Computer and Information Sciences, Pakistan Institute of Engineering and Applied Sciences, Nilore, 45650, Islamabad, Pakistan
| | - Shahzad Ahmad Qureshi
- Department of Computer and Information Sciences, Pakistan Institute of Engineering and Applied Sciences, Nilore, 45650, Islamabad, Pakistan
| | - Aziz Ul Rehman
- Agri & Biophotonics Division, National Institute of Lasers and Optronics College, Pakistan Institute of Engineering and Applied Sciences (PIEAS) P.O. Nilore, 45650 Islamabad, Pakistan.
| | | | - Seong-O Shim
- Faculty of Computing and IT, University of Jeddah, Jeddah, Saudi Arabia
| |
Collapse
|
48
|
A Novel Architecture to Classify Histopathology Images Using Convolutional Neural Networks. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10082929] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Histopathology is the study of tissue structure under the microscope to determine if the cells are normal or abnormal. Histopathology is a very important exam that is used to determine the patients’ treatment plan. The classification of histopathology images is very difficult to even an experienced pathologist, and a second opinion is often needed. Convolutional neural network (CNN), a particular type of deep learning architecture, obtained outstanding results in computer vision tasks like image classification. In this paper, we propose a novel CNN architecture to classify histopathology images. The proposed model consists of 15 convolution layers and two fully connected layers. A comparison between different activation functions was performed to detect the most efficient one, taking into account two different optimizers. To train and evaluate the proposed model, the publicly available PatchCamelyon dataset was used. The dataset consists of 220,000 annotated images for training and 57,000 unannotated images for testing. The proposed model achieved higher performance compared to the state-of-the-art architectures with an AUC of 95.46%.
Collapse
|
49
|
Bhattacharjee S, Kim CH, Park HG, Prakash D, Madusanka N, Cho NH, Choi HK. Multi-Features Classification of Prostate Carcinoma Observed in Histological Sections: Analysis of Wavelet-Based Texture and Colour Features. Cancers (Basel) 2019; 11:E1937. [PMID: 31817111 PMCID: PMC6966617 DOI: 10.3390/cancers11121937] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2019] [Revised: 11/14/2019] [Accepted: 11/28/2019] [Indexed: 11/16/2022] Open
Abstract
Microscopic biopsy images are coloured in nature because pathologists use the haematoxylin and eosin chemical colour dyes for biopsy examinations. In this study, biopsy images are used for histological grading and the analysis of benign and malignant prostate tissues. The following PCa grades are analysed in the present study: benign, grade 3, grade 4, and grade 5. Biopsy imaging has become increasingly important for the clinical assessment of PCa. In order to analyse and classify the histological grades of prostate carcinomas, pixel-based colour moment descriptor (PCMD) and gray-level co-occurrence matrix (GLCM) methods were used to extract the most significant features for multilayer perceptron (MLP) neural network classification. Haar wavelet transformation was carried out to extract GLCM texture features, and colour features were extracted from RGB (red/green/blue) colour images of prostate tissues. The MANOVA statistical test was performed to select significant features based on F-values and P-values using the R programming language. We obtained an average highest accuracy of 92.7% using level-1 wavelet texture and colour features. The MLP classifier performed well, and our study shows promising results based on multi-feature classification of histological sections of prostate carcinomas.
Collapse
Affiliation(s)
- Subrata Bhattacharjee
- Department of Computer Engineering, u-AHRC, Inje University, Gimhae 50834, Korea; (S.B.); (H.-G.P.); (D.P.); (N.M.)
| | - Cho-Hee Kim
- Department of Digital Anti-Aging Healthcare, Inje University, Gimhae 50834, Korea;
| | - Hyeon-Gyun Park
- Department of Computer Engineering, u-AHRC, Inje University, Gimhae 50834, Korea; (S.B.); (H.-G.P.); (D.P.); (N.M.)
| | - Deekshitha Prakash
- Department of Computer Engineering, u-AHRC, Inje University, Gimhae 50834, Korea; (S.B.); (H.-G.P.); (D.P.); (N.M.)
| | - Nuwan Madusanka
- Department of Computer Engineering, u-AHRC, Inje University, Gimhae 50834, Korea; (S.B.); (H.-G.P.); (D.P.); (N.M.)
| | - Nam-Hoon Cho
- Department of Pathology, Yonsei University Hospital, Seoul 03722, Korea;
| | - Heung-Kook Choi
- Department of Computer Engineering, u-AHRC, Inje University, Gimhae 50834, Korea; (S.B.); (H.-G.P.); (D.P.); (N.M.)
| |
Collapse
|
50
|
Automatic Detection of a Standard Line for Brain Magnetic Resonance Imaging Using Deep Learning. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9183849] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
Abstract
Recently, deep learning technology has been applied to medical images. This study aimed to create a detector able to automatically detect an anatomical structure presented in a brain magnetic resonance imaging (MRI) scan to draw a standard line. A total of 1200 brain sagittal MRI scans were used for training and validation. Two sizes of regions of interest (ROIs) were drawn on each anatomical structure measuring 64 × 64 pixels and 32 × 32 pixels, respectively. Data augmentation was applied to these ROIs. The faster region-based convolutional neural network was used as the network model for training. The detectors created were validated to evaluate the precision of detection. Anatomical structures detected by the model created were processed to draw the standard line. The average precision of anatomical detection, detection rate of the standard line, and accuracy rate of achieving a correct drawing were evaluated. For the 64 × 64-pixel ROI, the mean average precision achieved a result of 0.76 ± 0.04, which was higher than the outcome achieved with the 32 × 32-pixel ROI. Moreover, the detection and accuracy rates of the angle of difference at 10 degrees for the orbitomeatal line were 93.3 ± 5.2 and 76.7 ± 11.0, respectively. The automatic detection of a reference line for brain MRI can help technologists improve this examination.
Collapse
|