1
|
Lilhore UK, Dalal S, Faujdar N, Margala M, Chakrabarti P, Chakrabarti T, Simaiya S, Kumar P, Thangaraju P, Velmurugan H. Hybrid CNN-LSTM model with efficient hyperparameter tuning for prediction of Parkinson's disease. Sci Rep 2023; 13:14605. [PMID: 37669970 PMCID: PMC10480168 DOI: 10.1038/s41598-023-41314-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2023] [Accepted: 08/24/2023] [Indexed: 09/07/2023] Open
Abstract
The patients' vocal Parkinson's disease (PD) changes could be identified early on, allowing for management before physically incapacitating symptoms appear. In this work, static as well as dynamic speech characteristics that are relevant to PD identification are examined. Speech changes or communication issues are among the challenges that Parkinson's individuals may encounter. As a result, avoiding the potential consequences of speech difficulties brought on by the condition depends on getting the appropriate diagnosis early. PD patients' speech signals change significantly from those of healthy individuals. This research presents a hybrid model utilizing improved speech signals with dynamic feature breakdown using CNN and LSTM. The proposed hybrid model employs a new, pre-trained CNN with LSTM to recognize PD in linguistic features utilizing Mel-spectrograms derived from normalized voice signal and dynamic mode decomposition. The proposed Hybrid model works in various phases, which include Noise removal, extraction of Mel-spectrograms, feature extraction using pre-trained CNN model ResNet-50, and the final stage is applied for classification. An experimental analysis was performed using the PC-GITA disease dataset. The proposed hybrid model is compared with traditional NN and well-known machine learning-based CART and SVM & XGBoost models. The accuracy level achieved in Neural Network, CART, SVM, and XGBoost models is 72.69%, 84.21%, 73.51%, and 90.81%. The results show that under these four machine approaches of tenfold cross-validation and dataset splitting without samples overlapping one individual, the proposed hybrid model achieves an accuracy of 93.51%, significantly outperforming traditional ML models utilizing static features in detecting Parkinson's disease.
Collapse
Affiliation(s)
- Umesh Kumar Lilhore
- Department of Computer Science and Engineering, Chandigarh University, Chandigarh, Punjab, India
| | - Surjeet Dalal
- Amity School of Engineering and Technology, Amity University Haryana, Gurugram, India
| | - Neetu Faujdar
- Department of Computer Engineering and Application, GLA University, Mathura, Uttar Pradesh, India
| | - Martin Margala
- School of Computing and Informatics, University of Louisiana at Lafayette, Lafayette, USA
| | - Prasun Chakrabarti
- Department of Computer Science and Engineering, Sir Padampat Singhania University, Udaipur, 313601, Rajasthan, India
| | | | - Sarita Simaiya
- Department of Computer Science and Engineering, Chandigarh University, Chandigarh, Punjab, India
- Apex Institute of Technology, Department of Computer Science and Engineering, Chandigarh University, Mohali, Punjab, India
| | - Pawan Kumar
- Department of Computer Science and Engineering, Chandigarh University, Chandigarh, Punjab, India
- College of Computing Sciences & IT, Teerthanker Mahaveer University, Moradabad, Uttar Pradesh, India
| | - Pugazhenthan Thangaraju
- Department of Pharmacology, All India Institute of Medical Sciences, Raipur, Chhattisgarh, India.
| | - Hemasri Velmurugan
- Department of Pharmacology, All India Institute of Medical Sciences, Raipur, Chhattisgarh, India
| |
Collapse
|
2
|
Virk JS, Singh M, Singh M, Panjwani U, Ray K. A Multimodal Feature Fusion Framework for Sleep-Deprived Fatigue Detection to Prevent Accidents. SENSORS (BASEL, SWITZERLAND) 2023; 23:4129. [PMID: 37112470 PMCID: PMC10144633 DOI: 10.3390/s23084129] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 04/16/2023] [Accepted: 04/18/2023] [Indexed: 06/19/2023]
Abstract
Sleep-deprived fatigued person is likely to commit more errors that may even prove to be fatal. Thus, it is necessary to recognize this fatigue. The novelty of the proposed research work for the detection of this fatigue is that it is nonintrusive and based on multimodal feature fusion. In the proposed methodology, fatigue is detected by obtaining features from four domains: visual images, thermal images, keystroke dynamics, and voice features. In the proposed methodology, the samples of a volunteer (subject) are obtained from all four domains for feature extraction, and empirical weights are assigned to the four different domains. Young, healthy volunteers (n = 60) between the age group of 20 to 30 years participated in the experimental study. Further, they abstained from the consumption of alcohol, caffeine, or other drugs impacting their sleep pattern during the study. Through this multimodal technique, appropriate weights are given to the features obtained from the four domains. The results are compared with k-nearest neighbors (kNN), support vector machines (SVM), random tree, random forest, and multilayer perceptron classifiers. The proposed nonintrusive technique has obtained an average detection accuracy of 93.33% in 3-fold cross-validation.
Collapse
Affiliation(s)
- Jitender Singh Virk
- EIE Department, Thapar Institute of Engineering and Technology, Patiala 147001, India
| | - Mandeep Singh
- EIE Department, Thapar Institute of Engineering and Technology, Patiala 147001, India
| | - Mandeep Singh
- EIE Department, Thapar Institute of Engineering and Technology, Patiala 147001, India
| | - Usha Panjwani
- DIPAS, Defence Research and Development Organisation, Delhi 110054, India
| | - Koushik Ray
- DIPAS, Defence Research and Development Organisation, Delhi 110054, India
| |
Collapse
|
3
|
Zhang LM, Li Y, Zhang YT, Ng GW, Leau YB, Yan H. A Deep Learning Method Using Gender-Specific Features for Emotion Recognition. SENSORS (BASEL, SWITZERLAND) 2023; 23:1355. [PMID: 36772395 PMCID: PMC9921859 DOI: 10.3390/s23031355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/31/2022] [Revised: 01/20/2023] [Accepted: 01/22/2023] [Indexed: 06/18/2023]
Abstract
Speech reflects people's mental state and using a microphone sensor is a potential method for human-computer interaction. Speech recognition using this sensor is conducive to the diagnosis of mental illnesses. The gender difference of speakers affects the process of speech emotion recognition based on specific acoustic features, resulting in the decline of emotion recognition accuracy. Therefore, we believe that the accuracy of speech emotion recognition can be effectively improved by selecting different features of speech for emotion recognition based on the speech representations of different genders. In this paper, we propose a speech emotion recognition method based on gender classification. First, we use MLP to classify the original speech by gender. Second, based on the different acoustic features of male and female speech, we analyze the influence weights of multiple speech emotion features in male and female speech, and establish the optimal feature sets for male and female emotion recognition, respectively. Finally, we train and test CNN and BiLSTM, respectively, by using the male and the female speech emotion feature sets. The results show that the proposed emotion recognition models have an advantage in terms of average recognition accuracy compared with gender-mixed recognition models.
Collapse
Affiliation(s)
- Li-Min Zhang
- Key Laboratory for Artificial Intelligence and Cognitive Neuroscience of Language, Xi’an International Studies University, Xi’an 610116, China
- Faculty of Computing and Informatics, Universiti Malaysia Sabah, Sabah 88400, Malaysia
| | - Yang Li
- Key Laboratory for Artificial Intelligence and Cognitive Neuroscience of Language, Xi’an International Studies University, Xi’an 610116, China
| | - Yue-Ting Zhang
- Key Laboratory for Artificial Intelligence and Cognitive Neuroscience of Language, Xi’an International Studies University, Xi’an 610116, China
| | - Giap Weng Ng
- Faculty of Computing and Informatics, Universiti Malaysia Sabah, Sabah 88400, Malaysia
| | - Yu-Beng Leau
- Faculty of Computing and Informatics, Universiti Malaysia Sabah, Sabah 88400, Malaysia
| | - Hao Yan
- Key Laboratory for Artificial Intelligence and Cognitive Neuroscience of Language, Xi’an International Studies University, Xi’an 610116, China
| |
Collapse
|
4
|
Region-Based Segmentation and Classification for Ovarian Cancer Detection Using Convolution Neural Network. CONTRAST MEDIA & MOLECULAR IMAGING 2022; 2022:5968939. [PMID: 36475297 PMCID: PMC9701126 DOI: 10.1155/2022/5968939] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Revised: 07/15/2022] [Accepted: 07/18/2022] [Indexed: 11/21/2022]
Abstract
Ovarian cancer is a serious sickness for elderly women. According to data, it is the seventh leading cause of death in women as well as the fifth most frequent disease worldwide. Many researchers classified ovarian cancer using Artificial Neural Networks (ANNs). Doctors consider classification accuracy to be an important aspect of making decisions. Doctors consider improved classification accuracy for providing proper treatment. Early and precise diagnosis lowers mortality rates and saves lives. On basis of ROI (region of interest) segmentation, this research presents a novel annotated ovarian image classification utilizing FaRe-ConvNN (rapid region-based Convolutional neural network). The input photos were divided into three categories: epithelial, germ, and stroma cells. This image is segmented as well as preprocessed. After that, FaRe-ConvNN is used to perform the annotation procedure. For region-based classification, the method compares manually annotated features as well as trained feature in FaRe-ConvNN. This will aid in the analysis of higher accuracy in disease identification, as human annotation has lesser accuracy in previous studies; therefore, this effort will empirically prove that ML classification will provide higher accuracy. Classification is done using a combination of SVC and Gaussian NB classifiers after the region-based training in FaRe-ConvNN. The ensemble technique was employed in feature classification due to better data indexing. To diagnose ovarian cancer, the simulation provides an accurate portion of the input image. FaRe-ConvNN has a precision value of more than 95%, SVC has a precision value of 95.96%, and Gaussian NB has a precision value of 97.7%, with FR-CNN enhancing precision in Gaussian NB. For recall/sensitivity, SVC is 94.31 percent and Gaussian NB is 97.7 percent, while for specificity, SVC is 97.39 percent and Gaussian NB is 98.69 percent using FaRe-ConvNN.
Collapse
|
5
|
Analysis of Smart Lung Tumour Detector and Stage Classifier Using Deep Learning Techniques with Internet of Things. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:4608145. [PMID: 36148416 PMCID: PMC9489382 DOI: 10.1155/2022/4608145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/04/2022] [Revised: 07/15/2022] [Accepted: 07/27/2022] [Indexed: 11/30/2022]
Abstract
The use of artificial intelligence (AI) and the Internet of Things (IoT), which is a developing technology in medical applications that assists physicians in making more informed decisions regarding patients' courses of treatment, has become increasingly widespread in recent years in the field of healthcare. On the other hand, the number of PET scans that are being performed is rising, and radiologists are getting significantly overworked as a result. As a direct result of this, a novel approach that goes by the name “computer-aided diagnostics” is now being investigated as a potential method for reducing the tremendous workloads. A Smart Lung Tumor Detector and Stage Classifier (SLD-SC) is presented in this study as a hybrid technique for PET scans. This detector can identify the stage of a lung tumour. Following the development of the modified LSTM for the detection of lung tumours, the proposed SLD-SC went on to develop a Multilayer Convolutional Neural Network (M-CNN) for the classification of the various stages of lung cancer. This network was then modelled and validated utilising standard benchmark images. The suggested SLD-SC is now being evaluated on lung cancer pictures taken from patients with the disease. We observed that our recommended method gave good results when compared to other tactics that are currently being used in the literature. These findings were outstanding in terms of the performance metrics accuracy, recall, and precision that were assessed. As can be shown by the much better outcomes that were achieved with each of the test images that were used, our proposed method excels its rivals in a variety of respects. In addition to this, it achieves an average accuracy of 97 percent in the categorization of lung tumours, which is much higher than the accuracy achieved by the other approaches.
Collapse
|
6
|
Text-Based Emotion Recognition Using Deep Learning Approach. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:2645381. [PMID: 36052029 PMCID: PMC9427219 DOI: 10.1155/2022/2645381] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/29/2022] [Revised: 05/22/2022] [Accepted: 07/01/2022] [Indexed: 12/03/2022]
Abstract
Sentiment analysis is a method to identify people's attitudes, sentiments, and emotions towards a given goal, such as people, activities, organizations, services, subjects, and products. Emotion detection is a subset of sentiment analysis as it predicts the unique emotion rather than just stating positive, negative, or neutral. In recent times, many researchers have already worked on speech and facial expressions for emotion recognition. However, emotion detection in text is a tedious task as cues are missing, unlike in speech, such as tonal stress, facial expression, pitch, etc. To identify emotions from text, several methods have been proposed in the past using natural language processing (NLP) techniques: the keyword approach, the lexicon-based approach, and the machine learning approach. However, there were some limitations with keyword- and lexicon-based approaches as they focus on semantic relations. In this article, we have proposed a hybrid (machine learning + deep learning) model to identify emotions in text. Convolutional neural network (CNN) and Bi-GRU were exploited as deep learning techniques. Support vector machine is used as a machine learning approach. The performance of the proposed approach is evaluated using a combination of three different types of datasets, namely, sentences, tweets, and dialogs, and it attains an accuracy of 80.11%.
Collapse
|