1
|
Liver Disease Detection – Evaluation of Machine Learning Algorithms Performances with Optimal Thresholds. INTERNATIONAL JOURNAL OF HEALTHCARE INFORMATION SYSTEMS AND INFORMATICS 2022. [DOI: 10.4018/ijhisi.299956] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Intelligent predictive systems are showing a greater level of accuracy and effectiveness in early detection of critical diseases like cancer and liver and lung disease.Predictive models assist medical practitioners in identifying the diseases based on symptoms and health indicators like hormone,enzymes,age,bloodcounts,etc.This study proposes a framework to use classification models to accurately detect chronic liver disease by enhancing the prediction accuracy through cutting-edge analytics techniques.The article proposes an enhanced framework on the original study by Ramana et al. (2011).It uses evaluation measures like Precision and Balanced Accuracy to choose the most efficient classification algorithm in INDIA and USA patient datasets using various factors like enzymes,age,etc.Using Youden’s Index, individual thresholds for each model were identified to increase the power of sensitivity and specificity.A framework is proposed for highly accurate automated disease detection in the medical industry,and it helps in strategizing preventive measures for patients with liver diseases.
Collapse
|
2
|
Alizadehsani R, Roshanzamir M, Hussain S, Khosravi A, Koohestani A, Zangooei MH, Abdar M, Beykikhoshk A, Shoeibi A, Zare A, Panahiazar M, Nahavandi S, Srinivasan D, Atiya AF, Acharya UR. Handling of uncertainty in medical data using machine learning and probability theory techniques: a review of 30 years (1991-2020). ANNALS OF OPERATIONS RESEARCH 2021; 339:1-42. [PMID: 33776178 PMCID: PMC7982279 DOI: 10.1007/s10479-021-04006-2] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 02/23/2021] [Indexed: 05/17/2023]
Abstract
Understanding the data and reaching accurate conclusions are of paramount importance in the present era of big data. Machine learning and probability theory methods have been widely used for this purpose in various fields. One critically important yet less explored aspect is capturing and analyzing uncertainties in the data and model. Proper quantification of uncertainty helps to provide valuable information to obtain accurate diagnosis. This paper reviewed related studies conducted in the last 30 years (from 1991 to 2020) in handling uncertainties in medical data using probability theory and machine learning techniques. Medical data is more prone to uncertainty due to the presence of noise in the data. So, it is very important to have clean medical data without any noise to get accurate diagnosis. The sources of noise in the medical data need to be known to address this issue. Based on the medical data obtained by the physician, diagnosis of disease, and treatment plan are prescribed. Hence, the uncertainty is growing in healthcare and there is limited knowledge to address these problems. Our findings indicate that there are few challenges to be addressed in handling the uncertainty in medical raw data and new models. In this work, we have summarized various methods employed to overcome this problem. Nowadays, various novel deep learning techniques have been proposed to deal with such uncertainties and improve the performance in decision making.
Collapse
Affiliation(s)
- Roohallah Alizadehsani
- Institute for Intelligent Systems Research and Innovations (IISRI), Deakin University, Geelong, Australia
| | - Mohamad Roshanzamir
- Department of Computer Engineering, Faculty of Engineering, Fasa University, 74617-81189 Fasa, Iran
| | - Sadiq Hussain
- System Administrator, Dibrugarh University, Dibrugarh, Assam 786004 India
| | - Abbas Khosravi
- Institute for Intelligent Systems Research and Innovations (IISRI), Deakin University, Geelong, Australia
| | - Afsaneh Koohestani
- Institute for Intelligent Systems Research and Innovations (IISRI), Deakin University, Geelong, Australia
| | | | - Moloud Abdar
- Institute for Intelligent Systems Research and Innovations (IISRI), Deakin University, Geelong, Australia
| | - Adham Beykikhoshk
- Applied Artificial Intelligence Institute, Deakin University, Geelong, Australia
| | - Afshin Shoeibi
- Computer Engineering Department, Ferdowsi University of Mashhad, Mashhad, Iran
- Faculty of Electrical and Computer Engineering, Biomedical Data Acquisition Lab, K. N. Toosi University of Technology, Tehran, Iran
| | - Assef Zare
- Faculty of Electrical Engineering, Gonabad Branch, Islamic Azad University, Gonabad, Iran
| | - Maryam Panahiazar
- Institute for Computational Health Sciences, University of California, San Francisco, USA
| | - Saeid Nahavandi
- Institute for Intelligent Systems Research and Innovations (IISRI), Deakin University, Geelong, Australia
| | - Dipti Srinivasan
- Dept. of Electrical and Computer Engineering, National University of Singapore, Singapore, 117576 Singapore
| | - Amir F. Atiya
- Department of Computer Engineering, Faculty of Engineering, Cairo University, Cairo, 12613 Egypt
| | - U. Rajendra Acharya
- Department of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore, Singapore
- Department of Biomedical Engineering, School of Science and Technology, Singapore University of Social Sciences, Singapore, Singapore
- Department of Bioinformatics and Medical Engineering, Asia University, Taichung, Taiwan
| |
Collapse
|
4
|
Identification of the Framingham Risk Score by an Entropy-Based Rule Model for Cardiovascular Disease. ENTROPY 2020; 22:e22121406. [PMID: 33322122 PMCID: PMC7764435 DOI: 10.3390/e22121406] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/24/2020] [Revised: 11/30/2020] [Accepted: 12/11/2020] [Indexed: 12/12/2022]
Abstract
Since 2001, cardiovascular disease (CVD) has had the second-highest mortality rate, about 15,700 people per year, in Taiwan. It has thus imposed a substantial burden on medical resources. This study was triggered by the following three factors. First, the CVD problem reflects an urgent issue. A high priority has been placed on long-term therapy and prevention to reduce the wastage of medical resources, particularly in developed countries. Second, from the perspective of preventive medicine, popular data-mining methods have been well learned and studied, with excellent performance in medical fields. Thus, identification of the risk factors of CVD using these popular techniques is a prime concern. Third, the Framingham risk score is a core indicator that can be used to establish an effective prediction model to accurately diagnose CVD. Thus, this study proposes an integrated predictive model to organize five notable classifiers: the rough set (RS), decision tree (DT), random forest (RF), multilayer perceptron (MLP), and support vector machine (SVM), with a novel use of the Framingham risk score for attribute selection (i.e., F-attributes first identified in this study) to determine the key features for identifying CVD. Verification experiments were conducted with three evaluation criteria-accuracy, sensitivity, and specificity-based on 1190 instances of a CVD dataset available from a Taiwan teaching hospital and 2019 examples from a public Framingham dataset. Given the empirical results, the SVM showed the best performance in terms of accuracy (99.67%), sensitivity (99.93%), and specificity (99.71%) in all F-attributes in the CVD dataset compared to the other listed classifiers. The RS showed the highest performance in terms of accuracy (85.11%), sensitivity (86.06%), and specificity (85.19%) in most of the F-attributes in the Framingham dataset. The above study results support novel evidence that no classifier or model is suitable for all practical datasets of medical applications. Thus, identifying an appropriate classifier to address specific medical data is important. Significantly, this study is novel in its calculation and identification of the use of key Framingham risk attributes integrated with the DT technique to produce entropy-based decision rules of knowledge sets, which has not been undertaken in previous research. This study conclusively yielded meaningful entropy-based knowledgeable rules in tree structures and contributed to the differentiation of classifiers from the two datasets with three useful research findings and three helpful management implications for subsequent medical research. In particular, these rules provide reasonable solutions to simplify processes of preventive medicine by standardizing the formats and codes used in medical data to address CVD problems. The specificity of these rules is thus significant compared to those of past research.
Collapse
|
5
|
Taylor Bird Swarm Algorithm Based on Deep Belief Network for Heart Disease Diagnosis. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10186626] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Contemporary medicine depends on a huge amount of information contained in medical databases. Thus, the extraction of valuable knowledge, and making scientific decisions for the treatment of disease, has progressively become necessary to attain effective diagnosis. The obtainability of a large amount of medical data leads to the requirement of effective data analysis tools for extracting constructive knowledge. This paper proposes a novel method for heart disease diagnosis. Here, the pre-processing of medical data is done using log-transformation that converts the data to its uniform value range. Then, the feature selection process is performed using sparse fuzzy-c-means (FCM) for selecting significant features to classify medical data. Incorporating sparse FCM for the feature selection process provides more benefits for interpreting the models, as this sparse technique provides important features for detection, and can be utilized for handling high dimensional data. Then, the selected features are given to the deep belief network (DBN), which is trained using the proposed Taylor-based bird swarm algorithm (Taylor-BSA) for detection. Here, the proposed Taylor-BSA is designed by combining the Taylor series and bird swarm algorithm (BSA). The proposed Taylor-BSA–DBN outperformed other methods, with maximal accuracy of 93.4%, maximal sensitivity of 95%, and maximal specificity of 90.3%, respectively.
Collapse
|