1
|
Li L, Yang J, Por LY, Khan MS, Hamdaoui R, Hussain L, Iqbal Z, Rotaru IM, Dobrotă D, Aldrdery M, Omar A. Enhancing lung cancer detection through hybrid features and machine learning hyperparameters optimization techniques. Heliyon 2024; 10:e26192. [PMID: 38404820 PMCID: PMC10884486 DOI: 10.1016/j.heliyon.2024.e26192] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2023] [Revised: 01/30/2024] [Accepted: 02/08/2024] [Indexed: 02/27/2024] Open
Abstract
Machine learning offers significant potential for lung cancer detection, enabling early diagnosis and potentially improving patient outcomes. Feature extraction remains a crucial challenge in this domain. Combining the most relevant features can further enhance detection accuracy. This study employed a hybrid feature extraction approach, which integrates both Gray-level co-occurrence matrix (GLCM) with Haralick and autoencoder features with an autoencoder. These features were subsequently fed into supervised machine learning methods. Support Vector Machine (SVM) Radial Base Function (RBF) and SVM Gaussian achieved perfect performance measures, while SVM polynomial produced an accuracy of 99.89% when utilizing GLCM with an autoencoder, Haralick, and autoencoder features. SVM Gaussian achieved an accuracy of 99.56%, while SVM RBF achieved an accuracy of 99.35% when utilizing GLCM with Haralick features. These results demonstrate the potential of the proposed approach for developing improved diagnostic and prognostic lung cancer treatment planning and decision-making systems.
Collapse
Affiliation(s)
- Liangyu Li
- Center for Software Technology and Management, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, 43600, Bangi, Selangor, Malaysia
- Health Informatics Laboratory, Cancer Research Institute, Chifeng Cancer Hospital (Second Affiliated Hospital of Chifeng University), Medical Department, Chifeng University, Chifeng City, Inner Mongolia Autonomous Region, 024000, China
| | - Jing Yang
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603, Kuala Lumpur, Malaysia
| | - Lip Yee Por
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603, Kuala Lumpur, Malaysia
| | - Mohammad Shahbaz Khan
- Children's National Hospital, 111 Michigan Ave NW, Washington, DC, 20010, United States
| | - Rim Hamdaoui
- Department of Computer Science, College of Science and Human Studies Dawadmi, Shaqra University, Shaqra, Riyadh, Saudi Arabia
| | - Lal Hussain
- Department of Computer Science and Information Technology, King Abdullah Campus Chatter Kalas, University of Azad Jammu and Kashmir, Muzaffarabad, 13100, Azad Kashmir, Pakistan
- Department of Computer Science and Information Technology, Neelum Campus, University of Azad Jammu and Kashmir, Athmuqam, 13230, Azad Kashmir, Pakistan
| | - Zahoor Iqbal
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua, 321004, China
| | - Ionela Magdalena Rotaru
- Department of Industrial Engineering and Management, Lucian Blaga University of Sibiu, Bulevardul Victoriei 10, Sibiu, 550024, Romania
| | - Dan Dobrotă
- Faculty of Engineering, Lucian Blaga University of Sibiu, Bulevardul Victoriei 10, Sibiu, 550024, Romania
| | - Moutaz Aldrdery
- Department of Chemical Engineering, College of Engineering, King Khalid University, Abha, 61411, Saudi Arabia
| | - Abdulfattah Omar
- Department of English, College of Science & Humanities, Prince Sattam Bin Abdulaziz University, Saudi Arabia
| |
Collapse
|
2
|
Kumar S, Kumar H. Classification of COVID-19 X-ray images using transfer learning with visual geometrical groups and novel sequential convolutional neural networks. MethodsX 2023; 11:102295. [PMID: 37539339 PMCID: PMC10393783 DOI: 10.1016/j.mex.2023.102295] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2022] [Accepted: 07/20/2023] [Indexed: 08/05/2023] Open
Abstract
COVID-19 is a highly transmissible infectious disease that remains a substantial challenge. The utilization of chest radiology, particularly X-ray imaging, has proven to be highly effective, easily accessible, and cost-efficient in detecting COVID-19. A dataset named COVID-Xray-5k, consisting of imbalanced X-ray images of COVID-19-positive and normal subjects, is employed for investigation. The research introduces a novel methodology that utilizes conventional machine learning (ML), such as local binary patterns (LBP) for feature extraction and support vector machines (SVM) for classification. In addition, transfer learning is employed with the Visual Geometry Group 16-layer (VGG16) and 19-layer (VGG19) models. Besides, novel sequential convolutional neural network (CNN) architectures are presented to develop an autonomous system for classifying COVID-19. One of the proposed CNN architectures classifies the test dataset with an F1 score of 91.00% and an accuracy of 99.45% based on an empirical investigation to determine optimal hyper-parameters. The methods presented in the research show promising potential for COVID-19 classification, irrespective of class imbalance.•Employment of ML models to investigate subjective feature engineering and classification.•Transfer learning was employed for VGG16 and VGG19 with eight distinct models.•Illustration of two novel CNN sequential architectures; all the investigation is performed with and without weighted sampling.
Collapse
Affiliation(s)
- Sunil Kumar
- Department of Information Technology, School of Engineering and Technology (UIET), CSJM University, Kanpur, UP, India
- Department of Computer Engineering, J.C. Bose University of Science and Technology, YMCA, Faridabad, India
| | - Harish Kumar
- Department of Computer Engineering, J.C. Bose University of Science and Technology, YMCA, Faridabad, India
| |
Collapse
|
3
|
Shaheed K, Abbas Q, Hussain A, Qureshi I. Optimized Xception Learning Model and XgBoost Classifier for Detection of Multiclass Chest Disease from X-ray Images. Diagnostics (Basel) 2023; 13:2583. [PMID: 37568946 PMCID: PMC10416977 DOI: 10.3390/diagnostics13152583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Revised: 07/23/2023] [Accepted: 07/31/2023] [Indexed: 08/13/2023] Open
Abstract
Computed tomography (CT) scans, or radiographic images, were used to aid in the early diagnosis of patients and detect normal and abnormal lung function in the human chest. However, the diagnosis of lungs infected with coronavirus disease 2019 (COVID-19) was made more accurately from CT scan data than from a swab test. This study uses human chest radiography pictures to identify and categorize normal lungs, lung opacities, COVID-19-infected lungs, and viral pneumonia (often called pneumonia). In the past, several CAD systems using image processing, ML/DL, and other forms of machine learning have been developed. However, those CAD systems did not provide a general solution, required huge hyper-parameters, and were computationally inefficient to process huge datasets. Moreover, the DL models required high computational complexity, which requires a huge memory cost, and the complexity of the experimental materials' backgrounds, which makes it difficult to train an efficient model. To address these issues, we developed the Inception module, which was improved to recognize and detect four classes of Chest X-ray in this research by substituting the original convolutions with an architecture based on modified-Xception (m-Xception). In addition, the model incorporates depth-separable convolution layers within the convolution layer, interlinked by linear residuals. The model's training utilized a two-stage transfer learning process to produce an effective model. Finally, we used the XgBoost classifier to recognize multiple classes of chest X-rays. To evaluate the m-Xception model, the 1095 dataset was converted using a data augmentation technique into 48,000 X-ray images, including 12,000 normal, 12,000 pneumonia, 12,000 COVID-19 images, and 12,000 lung opacity images. To balance these classes, we used a data augmentation technique. Using public datasets with three distinct train-test divisions (80-20%, 70-30%, and 60-40%) to evaluate our work, we attained an average of 96.5% accuracy, 96% F1 score, 96% recall, and 96% precision. A comparative analysis demonstrates that the m-Xception method outperforms comparable existing methods. The results of the experiments indicate that the proposed approach is intended to assist radiologists in better diagnosing different lung diseases.
Collapse
Affiliation(s)
- Kashif Shaheed
- Department of Multimedia Systems, Faculty of Electronics, Telecommunication and Informatics, Gdansk University of Technology, 80-233 Gdansk, Poland;
| | - Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia;
| | - Ayyaz Hussain
- Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan;
| | - Imran Qureshi
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia;
| |
Collapse
|
4
|
Wu Y, Dravid A, Wehbe RM, Katsaggelos AK. DeepCOVID-Fuse: A Multi-Modality Deep Learning Model Fusing Chest X-rays and Clinical Variables to Predict COVID-19 Risk Levels. Bioengineering (Basel) 2023; 10:556. [PMID: 37237626 PMCID: PMC10215672 DOI: 10.3390/bioengineering10050556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2023] [Revised: 04/28/2023] [Accepted: 05/02/2023] [Indexed: 05/28/2023] Open
Abstract
The COVID-19 pandemic has posed unprecedented challenges to global healthcare systems, highlighting the need for accurate and timely risk prediction models that can prioritize patient care and allocate resources effectively. This study presents DeepCOVID-Fuse, a deep learning fusion model that predicts risk levels in patients with confirmed COVID-19 by combining chest radiographs (CXRs) and clinical variables. The study collected initial CXRs, clinical variables, and outcomes (i.e., mortality, intubation, hospital length of stay, Intensive care units (ICU) admission) from February to April 2020, with risk levels determined by the outcomes. The fusion model was trained on 1657 patients (Age: 58.30 ± 17.74; Female: 807) and validated on 428 patients (56.41 ± 17.03; 190) from the local healthcare system and tested on 439 patients (56.51 ± 17.78; 205) from a different holdout hospital. The performance of well-trained fusion models on full or partial modalities was compared using DeLong and McNemar tests. Results show that DeepCOVID-Fuse significantly (p < 0.05) outperformed models trained only on CXRs or clinical variables, with an accuracy of 0.658 and an area under the receiver operating characteristic curve (AUC) of 0.842. The fusion model achieves good outcome predictions even when only one of the modalities is used in testing, demonstrating its ability to learn better feature representations across different modalities during training.
Collapse
Affiliation(s)
- Yunan Wu
- Department of Electrical and Computer Engineering, Northwestern University, Evanston, IL 60201, USA;
| | - Amil Dravid
- Department of Computer Science, Northwestern University, Evanston, IL 60201, USA;
| | - Ramsey Michael Wehbe
- The Division of Cardiology, Department of Medicine and Bluhm Cardiovascular Institute, Northwestern Memorial Hospital, Chicago, IL 60611, USA;
| | - Aggelos K. Katsaggelos
- Department of Electrical and Computer Engineering, Northwestern University, Evanston, IL 60201, USA;
- Department of Computer Science, Northwestern University, Evanston, IL 60201, USA;
| |
Collapse
|