1
|
Malekroodi HS, Madusanka N, Lee BI, Yi M. Leveraging Deep Learning for Fine-Grained Categorization of Parkinson's Disease Progression Levels through Analysis of Vocal Acoustic Patterns. Bioengineering (Basel) 2024; 11:295. [PMID: 38534569 DOI: 10.3390/bioengineering11030295] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 03/18/2024] [Accepted: 03/18/2024] [Indexed: 03/28/2024] Open
Abstract
Speech impairments often emerge as one of the primary indicators of Parkinson's disease (PD), albeit not readily apparent in its early stages. While previous studies focused predominantly on binary PD detection, this research explored the use of deep learning models to automatically classify sustained vowel recordings into healthy controls, mild PD, or severe PD based on motor symptom severity scores. Popular convolutional neural network (CNN) architectures, VGG and ResNet, as well as vision transformers, Swin, were fine-tuned on log mel spectrogram image representations of the segmented voice data. Furthermore, the research investigated the effects of audio segment lengths and specific vowel sounds on the performance of these models. The findings indicated that implementing longer segments yielded better performance. The models showed strong capability in distinguishing PD from healthy subjects, achieving over 95% precision. However, reliably discriminating between mild and severe PD cases remained challenging. The VGG16 achieved the best overall classification performance with 91.8% accuracy and the largest area under the ROC curve. Furthermore, focusing analysis on the vowel /u/ could further improve accuracy to 96%. Applying visualization techniques like Grad-CAM also highlighted how CNN models focused on localized spectrogram regions while transformers attended to more widespread patterns. Overall, this work showed the potential of deep learning for non-invasive screening and monitoring of PD progression from voice recordings, but larger multi-class labeled datasets are needed to further improve severity classification.
Collapse
Affiliation(s)
- Hadi Sedigh Malekroodi
- Industry 4.0 Convergence Bionics Engineering, Pukyong National University, Busan 48513, Republic of Korea
| | - Nuwan Madusanka
- Digital of Healthcare Research Center, Institute of Information Technology and Convergence, Pukyong National University, Busan 48513, Republic of Korea
| | - Byeong-Il Lee
- Industry 4.0 Convergence Bionics Engineering, Pukyong National University, Busan 48513, Republic of Korea
- Digital of Healthcare Research Center, Institute of Information Technology and Convergence, Pukyong National University, Busan 48513, Republic of Korea
- Division of Smart Healthcare, Pukyong National University, Busan 48513, Republic of Korea
| | - Myunggi Yi
- Industry 4.0 Convergence Bionics Engineering, Pukyong National University, Busan 48513, Republic of Korea
- Digital of Healthcare Research Center, Institute of Information Technology and Convergence, Pukyong National University, Busan 48513, Republic of Korea
- Division of Smart Healthcare, Pukyong National University, Busan 48513, Republic of Korea
| |
Collapse
|
2
|
Khan A, Hassan T, Shafay M, Fahmy I, Werghi N, Mudigansalage S, Hussain I. Tomato maturity recognition with convolutional transformers. Sci Rep 2023; 13:22885. [PMID: 38129680 PMCID: PMC10739758 DOI: 10.1038/s41598-023-50129-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 12/15/2023] [Indexed: 12/23/2023] Open
Abstract
Tomatoes are a major crop worldwide, and accurately classifying their maturity is important for many agricultural applications, such as harvesting, grading, and quality control. In this paper, the authors propose a novel method for tomato maturity classification using a convolutional transformer. The convolutional transformer is a hybrid architecture that combines the strengths of convolutional neural networks (CNNs) and transformers. Additionally, this study introduces a new tomato dataset named KUTomaData, explicitly designed to train deep-learning models for tomato segmentation and classification. KUTomaData is a compilation of images sourced from a greenhouse in the UAE, with approximately 700 images available for training and testing. The dataset is prepared under various lighting conditions and viewing perspectives and employs different mobile camera sensors, distinguishing it from existing datasets. The contributions of this paper are threefold: firstly, the authors propose a novel method for tomato maturity classification using a modular convolutional transformer. Secondly, the authors introduce a new tomato image dataset that contains images of tomatoes at different maturity levels. Lastly, the authors show that the convolutional transformer outperforms state-of-the-art methods for tomato maturity classification. The effectiveness of the proposed framework in handling cluttered and occluded tomato instances was evaluated using two additional public datasets, Laboro Tomato and Rob2Pheno Annotated Tomato, as benchmarks. The evaluation results across these three datasets demonstrate the exceptional performance of our proposed framework, surpassing the state-of-the-art by 58.14%, 65.42%, and 66.39% in terms of mean average precision scores for KUTomaData, Laboro Tomato, and Rob2Pheno Annotated Tomato, respectively. This work can potentially improve the efficiency and accuracy of tomato harvesting, grading, and quality control processes.
Collapse
Affiliation(s)
- Asim Khan
- Department of Mechanical Engineering, Khalifa University, Abu Dhabi, UAE
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE
| | - Taimur Hassan
- Department of Electrical, Computer and Biomedical Engineering, Abu Dhabi University, Abu Dhabi, UAE
| | - Muhammad Shafay
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE
- Department of Electrical Engineering and Computer Science, Khalifa University, Abu Dhabi, UAE
| | - Israa Fahmy
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE
- Department of Electrical Engineering and Computer Science, Khalifa University, Abu Dhabi, UAE
| | - Naoufel Werghi
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE
- Department of Electrical Engineering and Computer Science, Khalifa University, Abu Dhabi, UAE
| | - Seneviratne Mudigansalage
- Department of Mechanical Engineering, Khalifa University, Abu Dhabi, UAE
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE
| | - Irfan Hussain
- Department of Mechanical Engineering, Khalifa University, Abu Dhabi, UAE.
- Khalifa University Center for Robotics and Autonomous Systems (KUCARS), Khalifa University, Abu Dhabi, UAE.
| |
Collapse
|
3
|
Tang H, Liang S, Yao D, Qiao Y. Improved STMask R-CNN-based defect detection model for automatic visual inspection of an optics lens. APPLIED OPTICS 2023; 62:8869-8881. [PMID: 38038033 DOI: 10.1364/ao.503039] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 10/31/2023] [Indexed: 12/02/2023]
Abstract
A lens defect is a common quality issue that has seriously harmed the scattering characteristics and performance of optical elements, reducing the quality consistency of the finished products. Furthermore, the energy hotspots coming from the high-energy laser through diffraction of optical component defects are amplified step by step in multi-level laser conduction, causing serious damage to the optical system. Traditional manual detection mainly relies on experienced workers under a special light source environment with high labor intensity, low efficiency, and accuracy. The common machine vision techniques are incapable of detecting low contrast and complex morphological defects. To address these challenges, a deep learning-based method, named STMask R-CNN, is proposed to detect defects on the surface and inside of a lens in complex environments. A Swin Transformer, which focuses on improving the modeling and representation capability of the features in order to improve the detection performance, is incorporated into the Mask R-CNN in this case. A challenge dataset containing more than 3800 images (18000 defect sample targets) with five different types of optical lens defects was created to verify the proposed approach. According to our experiments, the presented STMask R-CNN reached a precision value of 98.2%, recall value of 97.7%, F1 score of 97.9%, mAP@0.5 value of 98.1%, and FPS value of 24 f/s, which outperformed the SSD, Faster R-CNN, and YOLOv5. The experimental results demonstrated that the proposed STMask R-CNN outperformed other popular methods for multiscale targets, low contrast target detection and nesting, stacking, and intersecting defects sample detection, exhibiting good generalizability and robustness, as well as detection speed to meet mechanical equipment production efficiency requirements. In general, this research offers a favorable deep learning-based method for real-time automatic detection of optical lens defects.
Collapse
|
4
|
YOLOX-Dense-CT: a detection algorithm for cherry tomatoes based on YOLOX and DenseNet. JOURNAL OF FOOD MEASUREMENT AND CHARACTERIZATION 2022. [DOI: 10.1007/s11694-022-01553-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/15/2022]
|
5
|
Zheng H, Wang G, Li X. Identifying strawberry appearance quality by vision transformers and support vector machine. J FOOD PROCESS ENG 2022. [DOI: 10.1111/jfpe.14132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
- Hao Zheng
- School of Optoelectronic Engineering Xi'an Technological University Xi'an China
| | - Guohui Wang
- School of Optoelectronic Engineering Xi'an Technological University Xi'an China
| | - Xuchen Li
- School of Optoelectronic Engineering Xi'an Technological University Xi'an China
| |
Collapse
|