1
|
Alshammari K, Alshammari R, Alshammari A, Alkhudaydi T. An improved pear disease classification approach using cycle generative adversarial network. Sci Rep 2024; 14:6680. [PMID: 38509169 PMCID: PMC10955109 DOI: 10.1038/s41598-024-57143-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Accepted: 03/14/2024] [Indexed: 03/22/2024] Open
Abstract
A large number of countries worldwide depend on the agriculture, as agriculture can assist in reducing poverty, raising the country's income, and improving the food security. However, the plan diseases usually affect food crops and hence play a significant role in the annual yield and economic losses in the agricultural sector. In general, plant diseases have historically been identified by humans using their eyes, where this approach is often inexact, time-consuming, and exhausting. Recently, the employment of machine learning and deep learning approaches have significantly improved the classification and recognition accuracy for several applications. Despite the CNN models offer high accuracy for plant disease detection and classification, however, the limited available data for training the CNN model affects seriously the classification accuracy. Therefore, in this paper, we designed a Cycle Generative Adversarial Network (CycleGAN) to overcome the limitations of over-fitting and the limited size of the available datasets. In addition, we developed an efficient plant disease classification approach, where we adopt the CycleGAN architecture in order to enhance the classification accuracy. The obtained results showed an average enhancement of 7% in the classification accuracy.
Collapse
Affiliation(s)
- Khulud Alshammari
- Faculty of Computers & Information Technology, University of Tabuk, Tabuk, Saudi Arabia.
| | - Reem Alshammari
- Faculty of Computers & Information Technology, University of Tabuk, Tabuk, Saudi Arabia.
| | - Alanoud Alshammari
- Faculty of Computers & Information Technology, University of Tabuk, Tabuk, Saudi Arabia.
| | - Tahani Alkhudaydi
- Faculty of Computers & Information Technology, University of Tabuk, Tabuk, Saudi Arabia.
| |
Collapse
|
2
|
Abbas Q, Daadaa Y, Rashid U, Sajid MZ, Ibrahim MEA. HDR-EfficientNet: A Classification of Hypertensive and Diabetic Retinopathy Using Optimize EfficientNet Architecture. Diagnostics (Basel) 2023; 13:3236. [PMID: 37892058 PMCID: PMC10606674 DOI: 10.3390/diagnostics13203236] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Revised: 10/03/2023] [Accepted: 10/16/2023] [Indexed: 10/29/2023] Open
Abstract
Hypertensive retinopathy (HR) and diabetic retinopathy (DR) are retinal diseases closely associated with high blood pressure. The severity and duration of hypertension directly impact the prevalence of HR. The early identification and assessment of HR are crucial to preventing blindness. Currently, limited computer-aided methods are available for detecting HR and DR. These existing systems rely on traditional machine learning approaches, which require complex image processing techniques and are often limited in their application. To address this challenge, this work introduces a deep learning (DL) method called HDR-EfficientNet, which aims to provide an efficient and accurate approach to identifying various eye-related disorders, including diabetes and hypertensive retinopathy. The proposed method utilizes an EfficientNet-V2 network for end-to-end training focused on disease classification. Additionally, a spatial-channel attention method is incorporated into the approach to enhance its ability to identify specific areas of damage and differentiate between different illnesses. The HDR-EfficientNet model is developed using transfer learning, which helps overcome the challenge of imbalanced sample classes and improves the network's generalization. Dense layers are added to the model structure to enhance the feature selection capacity. The performance of the implemented system is evaluated using a large dataset of over 36,000 augmented retinal fundus images. The results demonstrate promising accuracy, with an average area under the curve (AUC) of 0.98, a specificity (SP) of 96%, an accuracy (ACC) of 98%, and a sensitivity (SE) of 95%. These findings indicate the effectiveness of the suggested HDR-EfficientNet classifier in diagnosing HR and DR. In summary, the HDR-EfficientNet method presents a DL-based approach that offers improved accuracy and efficiency for the detection and classification of HR and DR, providing valuable support in diagnosing and managing these eye-related conditions.
Collapse
Affiliation(s)
- Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (Y.D.)
| | - Yassine Daadaa
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (Y.D.)
| | - Umer Rashid
- Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan;
| | - Muhammad Zaheer Sajid
- Department of Computer Software Engineering, MCS, National University of Science and Technology, Islamabad 44000, Pakistan
| | - Mostafa E. A. Ibrahim
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (Y.D.)
- Department of Electrical Engineering, Benha Faculty of Engineering, Benha University, Benha 13518, Qalubia, Egypt
| |
Collapse
|
3
|
Heilmann PG, Frisch M, Abbadi A, Kox T, Herzog E. Stacked ensembles on basis of parentage information can predict hybrid performance with an accuracy comparable to marker-based GBLUP. FRONTIERS IN PLANT SCIENCE 2023; 14:1178902. [PMID: 37546247 PMCID: PMC10401275 DOI: 10.3389/fpls.2023.1178902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/03/2023] [Accepted: 06/26/2023] [Indexed: 08/08/2023]
Abstract
Testcross factorials in newly established hybrid breeding programs are often highly unbalanced, incomplete, and characterized by predominance of special combining ability (SCA) over general combining ability (GCA). This results in a low efficiency of GCA-based selection. Machine learning algorithms might improve prediction of hybrid performance in such testcross factorials, as they have been successfully applied to find complex underlying patterns in sparse data. Our objective was to compare the prediction accuracy of machine learning algorithms to that of GCA-based prediction and genomic best linear unbiased prediction (GBLUP) in six unbalanced incomplete factorials from hybrid breeding programs of rapeseed, wheat, and corn. We investigated a range of machine learning algorithms with three different types of predictor variables: (a) information on parentage of hybrids, (b) in addition hybrid performance of crosses of the parental lines with other crossing partners, and (c) genotypic marker data. In two highly incomplete and unbalanced factorials from rapeseed, in which the SCA variance contributed considerably to the genetic variance, stacked ensembles of gradient boosting machines based on parentage information outperformed GCA prediction. The stacked ensembles increased prediction accuracy from 0.39 to 0.45, and from 0.48 to 0.54 compared to GCA prediction. The prediction accuracy reached by stacked ensembles without marker data reached values comparable to those of GBLUP that requires marker data. We conclude that hybrid prediction with stacked ensembles of gradient boosting machines based on parentage information is a promising approach that is worth further investigations with other data sets in which SCA variance is high.
Collapse
Affiliation(s)
| | - Matthias Frisch
- Institute of Agronomy and Plant Breeding II, Justus Liebig University, Gießen, Germany
| | | | | | - Eva Herzog
- Institute of Agronomy and Plant Breeding II, Justus Liebig University, Gießen, Germany
| |
Collapse
|
4
|
Sajid MZ, Qureshi I, Abbas Q, Albathan M, Shaheed K, Youssef A, Ferdous S, Hussain A. Mobile-HR: An Ophthalmologic-Based Classification System for Diagnosis of Hypertensive Retinopathy Using Optimized MobileNet Architecture. Diagnostics (Basel) 2023; 13:diagnostics13081439. [PMID: 37189539 DOI: 10.3390/diagnostics13081439] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Revised: 04/13/2023] [Accepted: 04/15/2023] [Indexed: 05/17/2023] Open
Abstract
Hypertensive retinopathy (HR) is a serious eye disease that causes the retinal arteries to change. This change is mainly due to the fact of high blood pressure. Cotton wool patches, bleeding in the retina, and retinal artery constriction are affected lesions of HR symptoms. An ophthalmologist often makes the diagnosis of eye-related diseases by analyzing fundus images to identify the stages and symptoms of HR. The likelihood of vision loss can significantly decrease the initial detection of HR. In the past, a few computer-aided diagnostics (CADx) systems were developed to automatically detect HR eye-related diseases using machine learning (ML) and deep learning (DL) techniques. Compared to ML methods, the CADx systems use DL techniques that require the setting of hyperparameters, domain expert knowledge, a huge training dataset, and a high learning rate. Those CADx systems have shown to be good for automating the extraction of complex features, but they cause problems with class imbalance and overfitting. By ignoring the issues of a small dataset of HR, a high level of computational complexity, and the lack of lightweight feature descriptors, state-of-the-art efforts depend on performance enhancement. In this study, a pretrained transfer learning (TL)-based MobileNet architecture is developed by integrating dense blocks to optimize the network for the diagnosis of HR eye-related disease. We developed a lightweight HR-related eye disease diagnosis system, known as Mobile-HR, by integrating a pretrained model and dense blocks. To increase the size of the training and test datasets, we applied a data augmentation technique. The outcomes of the experiments show that the suggested approach was outperformed in many cases. This Mobile-HR system achieved an accuracy of 99% and an F1 score of 0.99 on different datasets. The results were verified by an expert ophthalmologist. These results indicate that the Mobile-HR CADx model produces positive outcomes and outperforms state-of-the-art HR systems in terms of accuracy.
Collapse
Affiliation(s)
- Muhammad Zaheer Sajid
- Department of Computer Software Engineering, MCS, National University of Science and Technology, Islamabad 44000, Pakistan
| | - Imran Qureshi
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
| | - Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
| | - Mubarak Albathan
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
| | - Kashif Shaheed
- Department of Multimedia Systems, Faculty of Electronics, Telecommunication and Informatics, Gdansk University of Technology, 80-233 Gdansk, Poland
| | - Ayman Youssef
- Department of Computers and Systems, Electronics Research Institute, Cairo 12622, Egypt
| | - Sehrish Ferdous
- Department of Software Engineering, National University of Modern Languages, Rawalpindi 44000, Pakistan
| | - Ayyaz Hussain
- Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan
| |
Collapse
|
5
|
Zhang Y, Aghajan ZM, Ison M, Lu Q, Tang H, Kalender G, Monsoor T, Zheng J, Kreiman G, Roychowdhury V, Fried I. Decoding of human identity by computer vision and neuronal vision. Sci Rep 2023; 13:651. [PMID: 36635322 PMCID: PMC9837190 DOI: 10.1038/s41598-022-26946-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2022] [Accepted: 12/22/2022] [Indexed: 01/14/2023] Open
Abstract
Extracting meaning from a dynamic and variable flow of incoming information is a major goal of both natural and artificial intelligence. Computer vision (CV) guided by deep learning (DL) has made significant strides in recognizing a specific identity despite highly variable attributes. This is the same challenge faced by the nervous system and partially addressed by the concept cells-neurons exhibiting selective firing in response to specific persons/places, described in the human medial temporal lobe (MTL) . Yet, access to neurons representing a particular concept is limited due to these neurons' sparse coding. It is conceivable, however, that the information required for such decoding is present in relatively small neuronal populations. To evaluate how well neuronal populations encode identity information in natural settings, we recorded neuronal activity from multiple brain regions of nine neurosurgical epilepsy patients implanted with depth electrodes, while the subjects watched an episode of the TV series "24". First, we devised a minimally supervised CV algorithm (with comparable performance against manually-labeled data) to detect the most prevalent characters (above 1% overall appearance) in each frame. Next, we implemented DL models that used the time-varying population neural data as inputs and decoded the visual presence of the four main characters throughout the episode. This methodology allowed us to compare "computer vision" with "neuronal vision"-footprints associated with each character present in the activity of a subset of neurons-and identify the brain regions that contributed to this decoding process. We then tested the DL models during a recognition memory task following movie viewing where subjects were asked to recognize clip segments from the presented episode. DL model activations were not only modulated by the presence of the corresponding characters but also by participants' subjective memory of whether they had seen the clip segment, and by the associative strengths of the characters in the narrative plot. The described approach can offer novel ways to probe the representation of concepts in time-evolving dynamic behavioral tasks. Further, the results suggest that the information required to robustly decode concepts is present in the population activity of only tens of neurons even in brain regions beyond MTL.
Collapse
Affiliation(s)
- Yipeng Zhang
- grid.19006.3e0000 0000 9632 6718Department of Electrical and Computer Engineering, University of California Los Angeles, Los Angeles, CA USA
| | - Zahra M. Aghajan
- grid.19006.3e0000 0000 9632 6718Department of Neurosurgery, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA USA
| | - Matias Ison
- grid.4563.40000 0004 1936 8868School of Psychology, University of Nottingham, Nottingham, UK
| | - Qiujing Lu
- grid.19006.3e0000 0000 9632 6718Department of Electrical and Computer Engineering, University of California Los Angeles, Los Angeles, CA USA
| | - Hanlin Tang
- grid.38142.3c000000041936754XChildren’s Hospital, Harvard Medical School, Boston, MA USA
| | - Guldamla Kalender
- grid.19006.3e0000 0000 9632 6718Department of Neurosurgery, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA USA
| | - Tonmoy Monsoor
- grid.19006.3e0000 0000 9632 6718Department of Electrical and Computer Engineering, University of California Los Angeles, Los Angeles, CA USA
| | - Jie Zheng
- grid.38142.3c000000041936754XChildren’s Hospital, Harvard Medical School, Boston, MA USA
| | - Gabriel Kreiman
- grid.38142.3c000000041936754XChildren’s Hospital, Harvard Medical School, Boston, MA USA ,grid.116068.80000 0001 2341 2786Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, MA USA
| | - Vwani Roychowdhury
- grid.19006.3e0000 0000 9632 6718Department of Electrical and Computer Engineering, University of California Los Angeles, Los Angeles, CA USA
| | - Itzhak Fried
- Department of Neurosurgery, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA, USA. .,Department of Psychiatry and Biobehavioral Sciences, Jane and Terry Semel Institute for Neuroscience and Human Behavior, University of California Los Angeles, Los Angeles, CA, USA. .,Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel.
| |
Collapse
|
6
|
Diwan T, Anirudh G, Tembhurne JV. Object detection using YOLO: challenges, architectural successors, datasets and applications. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 82:9243-9275. [PMID: 35968414 PMCID: PMC9358372 DOI: 10.1007/s11042-022-13644-y] [Citation(s) in RCA: 23] [Impact Index Per Article: 23.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Revised: 03/14/2022] [Accepted: 08/02/2022] [Indexed: 05/05/2023]
Abstract
Object detection is one of the predominant and challenging problems in computer vision. Over the decade, with the expeditious evolution of deep learning, researchers have extensively experimented and contributed in the performance enhancement of object detection and related tasks such as object classification, localization, and segmentation using underlying deep models. Broadly, object detectors are classified into two categories viz. two stage and single stage object detectors. Two stage detectors mainly focus on selective region proposals strategy via complex architecture; however, single stage detectors focus on all the spatial region proposals for the possible detection of objects via relatively simpler architecture in one shot. Performance of any object detector is evaluated through detection accuracy and inference time. Generally, the detection accuracy of two stage detectors outperforms single stage object detectors. However, the inference time of single stage detectors is better compared to its counterparts. Moreover, with the advent of YOLO (You Only Look Once) and its architectural successors, the detection accuracy is improving significantly and sometime it is better than two stage detectors. YOLOs are adopted in various applications majorly due to their faster inferences rather than considering detection accuracy. As an example, detection accuracies are 63.4 and 70 for YOLO and Fast-RCNN respectively, however, inference time is around 300 times faster in case of YOLO. In this paper, we present a comprehensive review of single stage object detectors specially YOLOs, regression formulation, their architecture advancements, and performance statistics. Moreover, we summarize the comparative illustration between two stage and single stage object detectors, among different versions of YOLOs, applications based on two stage detectors, and different versions of YOLOs along with the future research directions.
Collapse
Affiliation(s)
- Tausif Diwan
- Department of Computer Science & Engineering, Indian Institute of Information Technology, Nagpur, India
| | - G. Anirudh
- Department of Data science and analytics, Central University of Rajasthan, Jaipur, Rajasthan India
| | - Jitendra V. Tembhurne
- Department of Computer Science & Engineering, Indian Institute of Information Technology, Nagpur, India
| |
Collapse
|
7
|
Iwata H, Hayashi Y, Hasegawa A, Terayama K, Okuno Y. Classification of scanning electron microscope images of pharmaceutical excipients using deep convolutional neural networks with transfer learning. Int J Pharm X 2022; 4:100135. [PMID: 36325273 PMCID: PMC9619299 DOI: 10.1016/j.ijpx.2022.100135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Revised: 10/12/2022] [Accepted: 10/13/2022] [Indexed: 11/06/2022] Open
Abstract
Convolutional Neural Networks (CNNs) are image analysis techniques that have been applied to image classification in various fields. In this study, we applied a CNN to classify scanning electron microscopy (SEM) images of pharmaceutical raw material powders to determine if a CNN can evaluate particle morphology. We tested 10 pharmaceutical excipients with widely different particle morphologies. SEM images for each excipient were acquired and divided into training, validation, and test sets. Classification models were constructed by applying transfer learning to pretrained CNN models such as VGG16 and ResNet50. The results of a 5-fold cross-validation showed that the classification accuracy of the CNN model was sufficiently high using either pretrained model and that the type of excipient could be classified with high accuracy. The results suggest that the CNN model can detect differences in particle morphology, such as particle size, shape, and surface condition. By applying Grad-CAM to the constructed CNN model, we succeeded in finding particularly important regions in the particle image of the excipients. CNNs have been found to have the potential to be applied to the identification and characterization of raw material powders for pharmaceutical development.
Collapse
Affiliation(s)
- Hiroaki Iwata
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan
| | - Yoshihiro Hayashi
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan,Pharmaceutical Technology Division, Nichi-Iko Pharmaceutical Co., Ltd., 205-1, Shimoumezawa Namerikawa-shi, Toyama 936-0857, Japan,Correspondence to: Y. Hayashi, Pharmaceutical Technology Division, Nichi-Iko Pharmaceutical Co., Ltd.; 205-1, Shimoumezawa Namerikawa-shi, Toyama 936-0857, Japan.
| | - Aki Hasegawa
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan
| | - Kei Terayama
- Graduate School of Medical Life Science, Yokohama City University, Tsurumi-ku, Yokohama 230-0045, Japan
| | - Yasushi Okuno
- Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan,RIKEN Center for Computational Science, Kobe 650-0047, Japan,Correspondence to: Y. Okuno, Graduate School of Medicine, Kyoto University, 53 Shogoin-kawaharacho, Sakyo-ku, Kyoto 606-8507, Japan.
| |
Collapse
|
8
|
Tahir A, Munawar HS, Akram J, Adil M, Ali S, Kouzani AZ, Mahmud MAP. Automatic Target Detection from Satellite Imagery Using Machine Learning. SENSORS 2022; 22:s22031147. [PMID: 35161892 PMCID: PMC8839603 DOI: 10.3390/s22031147] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/29/2021] [Revised: 01/28/2022] [Accepted: 01/31/2022] [Indexed: 01/27/2023]
Abstract
Object detection is a vital step in satellite imagery-based computer vision applications such as precision agriculture, urban planning and defense applications. In satellite imagery, object detection is a very complicated task due to various reasons including low pixel resolution of objects and detection of small objects in the large scale (a single satellite image taken by Digital Globe comprises over 240 million pixels) satellite images. Object detection in satellite images has many challenges such as class variations, multiple objects pose, high variance in object size, illumination and a dense background. This study aims to compare the performance of existing deep learning algorithms for object detection in satellite imagery. We created the dataset of satellite imagery to perform object detection using convolutional neural network-based frameworks such as faster RCNN (faster region-based convolutional neural network), YOLO (you only look once), SSD (single-shot detector) and SIMRDWN (satellite imagery multiscale rapid detection with windowed networks). In addition to that, we also performed an analysis of these approaches in terms of accuracy and speed using the developed dataset of satellite imagery. The results showed that SIMRDWN has an accuracy of 97% on high-resolution images, while Faster RCNN has an accuracy of 95.31% on the standard resolution (1000 × 600). YOLOv3 has an accuracy of 94.20% on standard resolution (416 × 416) while on the other hand SSD has an accuracy of 84.61% on standard resolution (300 × 300). When it comes to speed and efficiency, YOLO is the obvious leader. In real-time surveillance, SIMRDWN fails. When YOLO takes 170 to 190 milliseconds to perform a task, SIMRDWN takes 5 to 103 milliseconds.
Collapse
Affiliation(s)
- Arsalan Tahir
- Research Center for Modeling and Simulation, National University of Sciences and Technology, Islamabad 64000, Pakistan; (A.T.); (M.A.)
| | - Hafiz Suliman Munawar
- School of Built Environment, University of New South Wales, Kensington, Sydney, NSW 2052, Australia
- Correspondence:
| | - Junaid Akram
- Department of Computer Science, Superior University, Lahore 54700, Pakistan; or
- School of Computer Science, The University of Sydney, Camperdown, Sydney, NSW 2006, Australia
| | - Muhammad Adil
- Research Center for Modeling and Simulation, National University of Sciences and Technology, Islamabad 64000, Pakistan; (A.T.); (M.A.)
| | - Shehryar Ali
- School of Engineering, Deakin University, Geelong, VIC 3216, Australia; (S.A.); (A.Z.K.); (M.A.P.M.)
| | - Abbas Z. Kouzani
- School of Engineering, Deakin University, Geelong, VIC 3216, Australia; (S.A.); (A.Z.K.); (M.A.P.M.)
| | - M. A. Pervez Mahmud
- School of Engineering, Deakin University, Geelong, VIC 3216, Australia; (S.A.); (A.Z.K.); (M.A.P.M.)
| |
Collapse
|
9
|
Alouani DJ, Ransom EM, Jani M, Burnham CA, Rhoads DD, Sadri N. Deep Convolutional Neural Networks Implementation for the Analysis of Urine Culture. Clin Chem 2022; 68:574-583. [PMID: 35134116 DOI: 10.1093/clinchem/hvab270] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2021] [Accepted: 11/12/2021] [Indexed: 12/14/2022]
Abstract
BACKGROUND Urine culture images collected using bacteriology automation are currently interpreted by technologists during routine standard-of-care workflows. Machine learning may be able to improve the harmonization of and assist with these interpretations. METHODS A deep learning model, BacterioSight, was developed, trained, and tested on standard BD-Kiestra images of routine blood agar urine cultures from 2 different medical centers. RESULTS BacterioSight displayed performance on par with standard-of-care-trained technologist interpretations. BacterioSight accuracy ranged from 97% when compared to standard-of-care (single technologist) and reached 100% when compared to a consensus reached by a group of technologists (gold standard in this study). Variability in image interpretation by trained technologists was identified and annotation "fuzziness" was quantified and found to correlate with reduced confidence in BacterioSight interpretation. Intra-testing (training and testing performed within the same institution) performed well giving Area Under the Curve (AUC) ≥0.98 for negative and positive plates, whereas, cross-testing on images (trained on one institution's images and tested on images from another institution) showed decreased performance with AUC ≥0.90 for negative and positive plates. CONCLUSIONS Our study provides a roadmap on how BacterioSight or similar deep learning prototypes may be implemented to screen for microbial growth, flag difficult cases for multi-personnel review, or auto-verify a subset of cultures with high confidence. In addition, our results highlight image interpretation variability by trained technologist within an institution and globally across institutions. We propose a model in which deep learning can enhance patient care by identifying inherent sample annotation variability and improving personnel training.
Collapse
Affiliation(s)
- David J Alouani
- Department of Pathology, University Hospitals Cleveland Medical Center, Cleveland, OH, USA
| | - Eric M Ransom
- Department of Pathology, University Hospitals Cleveland Medical Center, Cleveland, OH, USA
| | - Mehul Jani
- Department of Pathology, University Hospitals Cleveland Medical Center, Cleveland, OH, USA
| | - Carey-Ann Burnham
- Department of Pathology and Immunology, Washington University School of Medicine, St. Louis, MO, USA
| | - Daniel D Rhoads
- Department of Pathology, Cleveland Clinic Lerner College of Medicine, Cleveland, OH, USA
| | - Navid Sadri
- Department of Pathology, University Hospitals Cleveland Medical Center, Cleveland, OH, USA
| |
Collapse
|
10
|
Perez-Martin J, Bustos B, Guimarães SJF, Sipiran I, Pérez J, Said GC. A comprehensive review of the video-to-text problem. Artif Intell Rev 2022. [DOI: 10.1007/s10462-021-10104-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
11
|
Hassanzadeh T, Essam D, Sarker R. Evolutionary Deep Attention Convolutional Neural Networks for 2D and 3D Medical Image Segmentation. J Digit Imaging 2021; 34:1387-1404. [PMID: 34729668 PMCID: PMC8669068 DOI: 10.1007/s10278-021-00526-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2021] [Revised: 09/20/2021] [Accepted: 10/05/2021] [Indexed: 02/01/2023] Open
Abstract
Developing a convolutional neural network (CNN) for medical image segmentation is a complex task, especially when dealing with the limited number of available labelled medical images and computational resources. This task can be even more difficult if the aim is to develop a deep network and using a complicated structure like attention blocks. Because of various types of noises, artefacts and diversity in medical images, using complicated network structures like attention mechanism to improve the accuracy of segmentation is inevitable. Therefore, it is necessary to develop techniques to address the above difficulties. Neuroevolution is the combination of evolutionary computation and neural networks to establish a network automatically. However, Neuroevolution is computationally expensive, specifically to create 3D networks. In this paper, an automatic, efficient, accurate, and robust technique is introduced to develop deep attention convolutional neural networks utilising Neuroevolution for both 2D and 3D medical image segmentation. The proposed evolutionary technique can find a very good combination of six attention modules to recover spatial information from downsampling section and transfer them to the upsampling section of a U-Net-based network-six different CT and MRI datasets are employed to evaluate the proposed model for both 2D and 3D image segmentation. The obtained results are compared to state-of-the-art manual and automatic models, while our proposed model outperformed all of them.
Collapse
Affiliation(s)
| | - Daryl Essam
- University of New South Wales, Canberra, Australia
| | - Ruhul Sarker
- University of New South Wales, Canberra, Australia
| |
Collapse
|
12
|
Deep learning for object detection and scene perception in self-driving cars: Survey, challenges, and open issues. ARRAY 2021. [DOI: 10.1016/j.array.2021.100057] [Citation(s) in RCA: 53] [Impact Index Per Article: 17.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022] Open
|
13
|
Saini SK, Gupta R. Artificial intelligence methods for analysis of electrocardiogram signals for cardiac abnormalities: state-of-the-art and future challenges. Artif Intell Rev 2021. [DOI: 10.1007/s10462-021-09999-7] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
14
|
A Fuzzy Shell for Developing an Interpretable BCI Based on the Spatiotemporal Dynamics of the Evoked Oscillations. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2021; 2021:6685672. [PMID: 33936191 PMCID: PMC8055434 DOI: 10.1155/2021/6685672] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/17/2020] [Revised: 03/05/2021] [Accepted: 03/17/2021] [Indexed: 12/26/2022]
Abstract
Researchers in neuroscience computing experience difficulties when they try to carry out neuroanalysis in practice or when they need to design an explainable brain-computer interface (BCI) with quick setup and minimal training phase. There is a need of interpretable computational intelligence techniques and new brain states decoding for more understandable interpretation of the sensory, cognitive, and motor brain processing. We propose a general-purpose fuzzy software system shell for developing a custom EEG BCI system. It relies on the bursts of the ongoing EEG frequency power synchronization/desynchronization at scalp level and supports quick BCI setup by linguistic features, ad hoc fuzzy membership construction, explainable IF-THEN rules, and the concept of the Internet of Things (IoT), which makes the BCI system device and service independent. It has a potential for designing both passive and event-related BCIs with options for visual representation at scalp-source level in response to time. The feasibility of the proposed system has been proven by real experiments and bursts for β and γ frequency power have been detected in real time in response to evoked visuospatial selective attention. The presence of the proposed new brain state decoding can be used as a feasible metric for interpretation of the spatiotemporal dynamics of the passive or evoked neural oscillations.
Collapse
|
15
|
A Statistician Teaches Deep Learning. JOURNAL OF STATISTICAL THEORY AND PRACTICE 2021. [DOI: 10.1007/s42519-021-00193-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
16
|
Luo J, Lu K, Chen Y, Zhang B. Automatic identification of cashmere and wool fibers based on microscopic visual features and residual network model. Micron 2021; 143:103023. [PMID: 33540231 DOI: 10.1016/j.micron.2021.103023] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2020] [Accepted: 01/22/2021] [Indexed: 10/22/2022]
Abstract
Distinguishing cashmere and sheep wool fibers is a challenge. In this study, we propose a residual net-based method for the identification of cashmere and sheep wool fibers. First, optical microscopic images of six different types of cashmere and sheep wool fibers were collected, and then the sample images were data-augmented. Several classic convolutional neural network (CNN) models were trained and tested with the sample images. The comparison showed that the proposed residual net model with 18 weight layers had the highest accuracy, with an overall accuracy above 97.1 % on the test set; the highest accuracy on the Australian merino wool and Mongolian brown cashmere, both above 98 %; and the lowest accuracy on the Chinese white cashmere, above 95 %. The trained model exhibited a fast detection speed, processing 6000 sample images in less than 20 s.
Collapse
Affiliation(s)
- Junli Luo
- College of Information Engineering, Xuchang University, Xuchang, Henan, China; Xuchang Key Laboratory of Virtual Reality Technology, Xuchang, Henan, China
| | - Kai Lu
- College of Information Engineering, Xuchang University, Xuchang, Henan, China; Xuchang Key Laboratory of Virtual Reality Technology, Xuchang, Henan, China.
| | - Yonggang Chen
- College of Information Engineering, Xuchang University, Xuchang, Henan, China
| | - Boping Zhang
- College of Information Engineering, Xuchang University, Xuchang, Henan, China; Xuchang Key Laboratory of Virtual Reality Technology, Xuchang, Henan, China.
| |
Collapse
|
17
|
Abbas Q, Alsheddy A. Driver Fatigue Detection Systems Using Multi-Sensors, Smartphone, and Cloud-Based Computing Platforms: A Comparative Analysis. SENSORS (BASEL, SWITZERLAND) 2020; 21:E56. [PMID: 33374270 PMCID: PMC7796320 DOI: 10.3390/s21010056] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Revised: 12/17/2020] [Accepted: 12/20/2020] [Indexed: 12/16/2022]
Abstract
Internet of things (IoT) cloud-based applications deliver advanced solutions for smart cities to decrease traffic accidents caused by driver fatigue while driving on the road. Environmental conditions or driver behavior can ultimately lead to serious roadside accidents. In recent years, the authors have developed many low-cost, computerized, driver fatigue detection systems (DFDs) to help drivers, by using multi-sensors, and mobile and cloud-based computing architecture. To promote safe driving, these are the most current emerging platforms that were introduced in the past. In this paper, we reviewed state-of-the-art approaches for predicting unsafe driving styles using three common IoT-based architectures. The novelty of this article is to show major differences among multi-sensors, smartphone-based, and cloud-based architectures in multimodal feature processing. We discussed all of the problems that machine learning techniques faced in recent years, particularly the deep learning (DL) model, to predict driver hypovigilance, especially in terms of these three IoT-based architectures. Moreover, we performed state-of-the-art comparisons by using driving simulators to incorporate multimodal features of the driver. We also mention online data sources in this article to test and train network architecture in the field of DFDs on public available multimodal datasets. These comparisons assist other authors to continue future research in this domain. To evaluate the performance, we mention the major problems in these three architectures to help researchers use the best IoT-based architecture for detecting DFDs in a real-time environment. Moreover, the important factors of Multi-Access Edge Computing (MEC) and 5th generation (5G) networks are analyzed in the context of deep learning architecture to improve the response time of DFD systems. Lastly, it is concluded that there is a research gap when it comes to implementing the DFD systems on MEC and 5G technologies by using multimodal features and DL architecture.
Collapse
Affiliation(s)
- Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia;
| | | |
Collapse
|
18
|
Parida N, Mishra D, Das K, Rout NK, Panda G. On deep ensemble CNN–SAE based novel agro-market price forecasting. EVOLUTIONARY INTELLIGENCE 2020. [DOI: 10.1007/s12065-020-00466-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
19
|
Rashid M, Sulaiman N, P P Abdul Majeed A, Musa RM, Ab Nasir AF, Bari BS, Khatun S. Current Status, Challenges, and Possible Solutions of EEG-Based Brain-Computer Interface: A Comprehensive Review. Front Neurorobot 2020; 14:25. [PMID: 32581758 PMCID: PMC7283463 DOI: 10.3389/fnbot.2020.00025] [Citation(s) in RCA: 102] [Impact Index Per Article: 25.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2019] [Accepted: 04/08/2020] [Indexed: 12/12/2022] Open
Abstract
Brain-Computer Interface (BCI), in essence, aims at controlling different assistive devices through the utilization of brain waves. It is worth noting that the application of BCI is not limited to medical applications, and hence, the research in this field has gained due attention. Moreover, the significant number of related publications over the past two decades further indicates the consistent improvements and breakthroughs that have been made in this particular field. Nonetheless, it is also worth mentioning that with these improvements, new challenges are constantly discovered. This article provides a comprehensive review of the state-of-the-art of a complete BCI system. First, a brief overview of electroencephalogram (EEG)-based BCI systems is given. Secondly, a considerable number of popular BCI applications are reviewed in terms of electrophysiological control signals, feature extraction, classification algorithms, and performance evaluation metrics. Finally, the challenges to the recent BCI systems are discussed, and possible solutions to mitigate the issues are recommended.
Collapse
Affiliation(s)
- Mamunur Rashid
- Faculty of Electrical & Electronics Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| | - Norizam Sulaiman
- Faculty of Electrical & Electronics Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| | - Anwar P P Abdul Majeed
- Innovative Manufacturing, Mechatronics and Sports Laboratory, Faculty of Manufacturing and Mechatronic Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| | - Rabiu Muazu Musa
- Centre for Fundamental and Continuing Education, Universiti Malaysia Terengganu, Kuala Nerus, Malaysia
| | - Ahmad Fakhri Ab Nasir
- Innovative Manufacturing, Mechatronics and Sports Laboratory, Faculty of Manufacturing and Mechatronic Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| | - Bifta Sama Bari
- Faculty of Electrical & Electronics Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| | - Sabira Khatun
- Faculty of Electrical & Electronics Engineering Technology, Universiti Malaysia Pahang, Pekan, Malaysia
| |
Collapse
|
20
|
Khan A, Sohail A, Zahoora U, Qureshi AS. A survey of the recent architectures of deep convolutional neural networks. Artif Intell Rev 2020. [DOI: 10.1007/s10462-020-09825-6] [Citation(s) in RCA: 351] [Impact Index Per Article: 87.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
|
21
|
|