1
|
Zhang B, Zhao D. An Ensemble Learning Model for Detecting Soybean Seedling Emergence in UAV Imagery. SENSORS (BASEL, SWITZERLAND) 2023; 23:6662. [PMID: 37571446 PMCID: PMC10422598 DOI: 10.3390/s23156662] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/20/2023] [Revised: 07/19/2023] [Accepted: 07/24/2023] [Indexed: 08/13/2023]
Abstract
Efficient detection and evaluation of soybean seedling emergence is an important measure for making field management decisions. However, there are many indicators related to emergence, and using multiple models to detect them separately makes data processing too slow to aid timely field management. In this study, we aimed to integrate several deep learning and image processing methods to build a model to evaluate multiple soybean seedling emergence information. An unmanned aerial vehicle (UAV) was used to acquire soybean seedling RGB images at emergence (VE), cotyledon (VC), and first node (V1) stages. The number of soybean seedlings that emerged was obtained by the seedling emergence detection module, and image datasets were constructed using the seedling automatic cutting module. The improved AlexNet was used as the backbone network of the growth stage discrimination module. The above modules were combined to calculate the emergence proportion in each stage and determine soybean seedlings emergence uniformity. The results show that the seedling emergence detection module was able to identify the number of soybean seedlings with an average accuracy of 99.92%, a R2 of 0.9784, a RMSE of 6.07, and a MAE of 5.60. The improved AlexNet was more lightweight, training time was reduced, the average accuracy was 99.07%, and the average loss was 0.0355. The model was validated in the field, and the error between predicted and real emergence proportions was up to 0.0775 and down to 0.0060. It provides an effective ensemble learning model for the detection and evaluation of soybean seedling emergence, which can provide a theoretical basis for making decisions on soybean field management and precision operations and has the potential to evaluate other crops emergence information.
Collapse
Affiliation(s)
- Bo Zhang
- College of Information and Electrical Engineering, Heilongjiang Bayi Agricultural University, Daqing 163319, China
| | | |
Collapse
|
2
|
Alajlan NN, Ibrahim DM. DDD TinyML: A TinyML-Based Driver Drowsiness Detection Model Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2023; 23:5696. [PMID: 37420860 DOI: 10.3390/s23125696] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Revised: 06/07/2023] [Accepted: 06/14/2023] [Indexed: 07/09/2023]
Abstract
Driver drowsiness is one of the main causes of traffic accidents today. In recent years, driver drowsiness detection has suffered from issues integrating deep learning (DL) with Internet-of-things (IoT) devices due to the limited resources of IoT devices, which pose a challenge to fulfilling DL models that demand large storage and computation. Thus, there are challenges to meeting the requirements of real-time driver drowsiness detection applications that need short latency and lightweight computation. To this end, we applied Tiny Machine Learning (TinyML) to a driver drowsiness detection case study. In this paper, we first present an overview of TinyML. After conducting some preliminary experiments, we proposed five lightweight DL models that can be deployed on a microcontroller. We applied three DL models: SqueezeNet, AlexNet, and CNN. In addition, we adopted two pretrained models (MobileNet-V2 and MobileNet-V3) to find the best model in terms of size and accuracy results. After that, we applied the optimization methods to DL models using quantization. Three quantization methods were applied: quantization-aware training (QAT), full-integer quantization (FIQ), and dynamic range quantization (DRQ). The obtained results in terms of the model size show that the CNN model achieved the smallest size of 0.05 MB using the DRQ method, followed by SqueezeNet, AlexNet MobileNet-V3, and MobileNet-V2, with 0.141 MB, 0.58 MB, 1.16 MB, and 1.55 MB, respectively. The result after applying the optimization method was 0.9964 accuracy using DRQ in the MobileNet-V2 model, which outperformed the other models, followed by the SqueezeNet and AlexNet models, with 0.9951 and 0.9924 accuracies, respectively, using DRQ.
Collapse
Affiliation(s)
- Norah N Alajlan
- Department of Information Technology, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
| | - Dina M Ibrahim
- Department of Information Technology, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
- Department of Computers and Control Engineering, Faculty of Engineering, Tanta University, Tanta 31733, Egypt
| |
Collapse
|
3
|
Wei Z, Dong S, Wang X. Petrochemical Equipment Tracking by Improved Yolov7 Network and Hybrid Matching in Moving Scenes. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23094546. [PMID: 37177751 PMCID: PMC10181657 DOI: 10.3390/s23094546] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Revised: 04/22/2023] [Accepted: 05/02/2023] [Indexed: 05/15/2023]
Abstract
Petrochemical equipment tracking is a fundamental and important technology in petrochemical industry security monitoring, equipment working risk analysis, and other applications. In complex scenes where the multiple pipelines present different directions and many kinds of equipment have huge scale and shape variation in seriously mutual occlusions captured by moving cameras, the accuracy and speed of petrochemical equipment tracking would be limited because of the false and missed tracking of equipment with extreme sizes and severe occlusion, due to image quality, equipment scale, light, and other factors. In this paper, a new multiple petrochemical equipment tracking method is proposed by combining an improved Yolov7 network with attention mechanism and small target perceive layer and a hybrid matching that incorporates deep feature and traditional texture and location feature. The model incorporates the advantages of channel and spatial attention module into the improved Yolov7 detector and Siamese neural network for similarity matching. The proposed model is validated on the self-built petrochemical equipment video data set and the experimental results show it achieves a competitive performance in comparison with the related state-of-the-art tracking algorithms.
Collapse
Affiliation(s)
- Zhenqiang Wei
- College of Safety and Ocean Engineering, China University of Petroleum, Beijing 102249, China
- CNPC Research Institute of Safety & Environment Technology, Beijing 102206, China
| | - Shaohua Dong
- College of Safety and Ocean Engineering, China University of Petroleum, Beijing 102249, China
| | - Xuchu Wang
- Key Laboratory of Optoelectronic Technology and Systems of Ministry of Education, Chongqing University, Chongqing 400040, China
- College of Optoelectronic Engineering, Chongqing University, Chongqing 400040, China
| |
Collapse
|
4
|
Harimi A, Majd Y, Gharahbagh AA, Hajihashemi V, Esmaileyan Z, Machado JJM, Tavares JMRS. Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning. SENSORS (BASEL, SWITZERLAND) 2022; 22:9569. [PMID: 36559937 PMCID: PMC9782852 DOI: 10.3390/s22249569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/12/2022] [Revised: 12/04/2022] [Accepted: 12/04/2022] [Indexed: 06/17/2023]
Abstract
Heart sounds convey important information regarding potential heart diseases. Currently, heart sound classification attracts many researchers from the fields of telemedicine, digital signal processing, and machine learning-among others-mainly to identify cardiac pathology as quickly as possible. This article proposes chaogram as a new transform to convert heart sound signals to colour images. In the proposed approach, the output image is, therefore, the projection of the reconstructed phase space representation of the phonocardiogram (PCG) signal on three coordinate planes. This has two major benefits: (1) it makes possible to apply deep convolutional neural networks to heart sounds and (2) it is also possible to employ a transfer learning scheme by converting a heart sound signal to an image. The performance of the proposed approach was verified on the PhysioNet dataset. Due to the imbalanced data on this dataset, it is common to assess the results quality using the average of sensitivity and specificity, which is known as score, instead of accuracy. In this study, the best results were achieved using the InceptionV3 model, which achieved a score of 88.06%.
Collapse
Affiliation(s)
- Ali Harimi
- Department of Electrical Engineering, Shahrood Branch, Islamic Azad University, Shahrood 43189-36199, Iran
| | - Yahya Majd
- School of Surveying and Built Environment, Toowoomba Campus, University of Southern Queensland (USQ), Darling Heights, QLD 4350, Australia
| | | | - Vahid Hajihashemi
- Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| | - Zeynab Esmaileyan
- Department of Electrical Engineering, Shahrood Branch, Islamic Azad University, Shahrood 43189-36199, Iran
| | - José J. M. Machado
- Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| | - João Manuel R. S. Tavares
- Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| |
Collapse
|
5
|
Zhu J, Zeng Q, Han F, Cao H, Bian Y, Wei C. Study on the Construction of a Time-Space Four-Dimensional Combined Imaging Model and Moving Target Location Prediction Model. SENSORS (BASEL, SWITZERLAND) 2022; 22:6375. [PMID: 36080834 PMCID: PMC9459777 DOI: 10.3390/s22176375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/09/2022] [Revised: 08/16/2022] [Accepted: 08/19/2022] [Indexed: 06/15/2023]
Abstract
Time-space four-dimensional motion target localization is a fundamental and challenging task in the field of intelligent driving, and an important part of achieving the upgrade in existing target localization technologies. In order to solve the problem of the lack of localization of moving targets in a spatio-temporal four-dimensional environment in the existing spatio-temporal data model, this paper proposes an optical imaging model in the four-dimensional time-space system and a mathematical model of the object-image point mapping relationship in the four-dimensional time-space system based on the central perspective projection model, combined with the one-dimensional "time" and three-dimensional "space". After adding the temporal dimension, the imaging system parameters are extended. In order to solve the nonlinear mapping problem of complex systems, this paper proposes to construct a time-space four-dimensional object-image mapping relationship model based on a BP artificial neural network and demonstrates the feasibility of the joint time-space four-dimensional imaging model theory. In addition, indoor time-space four-dimensional localization prediction experiments verify the performance of the model in this paper. The maximum relative error rates of the predicted motion depth values, time values, and velocity values of this localization method compared with the real values do not exceed 0.23%, 2.03%, and 1.51%, respectively.
Collapse
Affiliation(s)
- Junchao Zhu
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
- Tianjin Key Laboratory for Control Theory and Applications in Complicated Systems, Tianjin University of Technology, Tianjin 300384, China
| | - Qi Zeng
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
- Tianjin Key Laboratory for Control Theory and Applications in Complicated Systems, Tianjin University of Technology, Tianjin 300384, China
| | - Fangfang Han
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
- Tianjin Key Laboratory for Control Theory and Applications in Complicated Systems, Tianjin University of Technology, Tianjin 300384, China
| | - Huifeng Cao
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
- Tianjin Key Laboratory for Control Theory and Applications in Complicated Systems, Tianjin University of Technology, Tianjin 300384, China
| | - Yongxin Bian
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
- Tianjin Key Laboratory for Control Theory and Applications in Complicated Systems, Tianjin University of Technology, Tianjin 300384, China
| | - Chenhong Wei
- School of Electrical Engineering and Automation, Tianjin University of Technology, Tianjin 300384, China
| |
Collapse
|
6
|
Predicting Activity Duration in Smart Sensing Environments Using Synthetic Data and Partial Least Squares Regression: The Case of Dementia Patients. SENSORS 2022; 22:s22145410. [PMID: 35891090 PMCID: PMC9318990 DOI: 10.3390/s22145410] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Revised: 07/11/2022] [Accepted: 07/13/2022] [Indexed: 12/10/2022]
Abstract
The accurate recognition of activities is fundamental for following up on the health progress of people with dementia (PwD), thereby supporting subsequent diagnosis and treatments. When monitoring the activities of daily living (ADLs), it is feasible to detect behaviour patterns, parse out the disease evolution, and consequently provide effective and timely assistance. However, this task is affected by uncertainties derived from the differences in smart home configurations and the way in which each person undertakes the ADLs. One adjacent pathway is to train a supervised classification algorithm using large-sized datasets; nonetheless, obtaining real-world data is costly and characterized by a challenging recruiting research process. The resulting activity data is then small and may not capture each person’s intrinsic properties. Simulation approaches have risen as an alternative efficient choice, but synthetic data can be significantly dissimilar compared to real data. Hence, this paper proposes the application of Partial Least Squares Regression (PLSR) to approximate the real activity duration of various ADLs based on synthetic observations. First, the real activity duration of each ADL is initially contrasted with the one derived from an intelligent environment simulator. Following this, different PLSR models were evaluated for estimating real activity duration based on synthetic variables. A case study including eight ADLs was considered to validate the proposed approach. The results revealed that simulated and real observations are significantly different in some ADLs (p-value < 0.05), nevertheless synthetic variables can be further modified to predict the real activity duration with high accuracy (R2(pred)>90%).
Collapse
|
7
|
Fakieh B, AL-Ghamdi ASALM, Ragab M. Optimal Deep Stacked Sparse Autoencoder Based Osteosarcoma Detection and Classification Model. Healthcare (Basel) 2022; 10:1040. [PMID: 35742091 PMCID: PMC9222514 DOI: 10.3390/healthcare10061040] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2022] [Revised: 05/30/2022] [Accepted: 05/30/2022] [Indexed: 02/04/2023] Open
Abstract
Osteosarcoma is a kind of bone cancer which generally starts to develop in the lengthy bones in the legs and arms. Because of an increase in occurrence of cancer and patient-specific treatment options, the detection and classification of cancer becomes a difficult process. The manual recognition of osteosarcoma necessitates expert knowledge and is time consuming. An earlier identification of osteosarcoma can reduce the death rate. With the development of new technologies, automated detection models can be exploited for medical image classification, thereby decreasing the expert's reliance and resulting in timely identification. In recent times, an amount of Computer-Aided Detection (CAD) systems are available in the literature for the segmentation and detection of osteosarcoma using medicinal images. In this view, this research work develops a wind driven optimization with deep transfer learning enabled osteosarcoma detection and classification (WDODTL-ODC) method. The presented WDODTL-ODC model intends to determine the presence of osteosarcoma in the biomedical images. To accomplish this, the osteosarcoma model involves Gaussian filtering (GF) based on pre-processing and contrast enhancement techniques. In addition, deep transfer learning using a SqueezNet model is utilized as a featured extractor. At last, the Wind Driven Optimization (WDO) algorithm with a deep-stacked sparse auto-encoder (DSSAE) is employed for the classification process. The simulation outcome demonstrated that the WDODTL-ODC technique outperformed the existing models in the detection of osteosarcoma on biomedical images.
Collapse
Affiliation(s)
- Bahjat Fakieh
- Information Systems Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia; (B.F.); (A.S.A.-M.A.-G.)
| | - Abdullah S. AL-Malaise AL-Ghamdi
- Information Systems Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia; (B.F.); (A.S.A.-M.A.-G.)
- Information Systems Department, HECI School, Dar Alhekma University, Jeddah 22246, Saudi Arabia
- Center of Excellence in Smart Environment Research, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| | - Mahmoud Ragab
- Center of Excellence in Smart Environment Research, King Abdulaziz University, Jeddah 21589, Saudi Arabia
- Information Technology Department, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
- Department of Mathematics, Faculty of Science, Al-Azhar University, Naser City, Cairo 11884, Egypt
- Centre for Artificial Intelligence in Precision Medicines, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| |
Collapse
|
8
|
Krishnan AM, Bouazizi M, Ohtsuki T. An Infrared Array Sensor-Based Approach for Activity Detection, Combining Low-Cost Technology with Advanced Deep Learning Techniques. SENSORS 2022; 22:s22103898. [PMID: 35632305 PMCID: PMC9145665 DOI: 10.3390/s22103898] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/26/2022] [Revised: 05/17/2022] [Accepted: 05/18/2022] [Indexed: 11/16/2022]
Abstract
In this paper, we propose an activity detection system using a 24 × 32 resolution infrared array sensor placed on the ceiling. We first collect the data at different resolutions (i.e., 24 × 32, 12 × 16, and 6 × 8) and apply the advanced deep learning (DL) techniques of Super-Resolution (SR) and denoising to enhance the quality of the images. We then classify the images/sequences of images depending on the activities the subject is performing using a hybrid deep learning model combining a Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM). We use data augmentation to improve the training of the neural networks by incorporating a wider variety of samples. The process of data augmentation is performed by a Conditional Generative Adversarial Network (CGAN). By enhancing the images using SR, removing the noise, and adding more training samples via data augmentation, our target is to improve the classification accuracy of the neural network. Through experiments, we show that employing these deep learning techniques to low-resolution noisy infrared images leads to a noticeable improvement in performance. The classification accuracy improved from 78.32% to 84.43% (for images with 6 × 8 resolution), and from 90.11% to 94.54% (for images with 12 × 16 resolution) when we used the CNN and CNN + LSTM networks, respectively.
Collapse
Affiliation(s)
| | - Mondher Bouazizi
- Faculty of Science and Technology, Keio University, Yokohama 223-8522, Japan;
| | - Tomoaki Ohtsuki
- Faculty of Science and Technology, Keio University, Yokohama 223-8522, Japan;
- Correspondence:
| |
Collapse
|