1
|
Shah A, Attique Khan M, Ibrahim Alzahrani A, Alalwan N, Hamza A, Manic S, Zhang Y, Damaševic̆ius R. FuzzyShallow: A framework of deep shallow neural networks and modified tree growth optimization for agriculture land cover and fruit disease recognition from remote sensing and digital imaging. MEASUREMENT 2024; 237:115224. [DOI: 10.1016/j.measurement.2024.115224] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/25/2024]
|
2
|
Cheng C, Liu W, Feng L, Jia Z. Emotion recognition using hierarchical spatial-temporal learning transformer from regional to global brain. Neural Netw 2024; 179:106624. [PMID: 39163821 DOI: 10.1016/j.neunet.2024.106624] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2024] [Revised: 06/10/2024] [Accepted: 08/09/2024] [Indexed: 08/22/2024]
Abstract
Emotion recognition is an essential but challenging task in human-computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial-temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial-temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial-temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the long-range spatial-temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.
Collapse
Affiliation(s)
- Cheng Cheng
- Department of Computer Science and Technology, Dalian University of Technology, Dalian, China
| | - Wenzhe Liu
- School of Information Engineering, Huzhou University, Huzhou, China
| | - Lin Feng
- Department of Computer Science and Technology, Dalian University of Technology, Dalian, China; School of Information and Communication Engineering, Dalian Minzu University, Dlian, China.
| | - Ziyu Jia
- Brainnetome Center, Institute of Automation, Chinese Academy of Sciences, University of Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
3
|
Kardoost A, Schönherr R, Deiter C, Redecke L, Lorenzen K, Schulz J, de Diego I. Convolutional neural network approach for the automated identification of in cellulo crystals. J Appl Crystallogr 2024; 57:266-275. [PMID: 38596734 PMCID: PMC11001417 DOI: 10.1107/s1600576724000682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Accepted: 01/18/2024] [Indexed: 04/11/2024] Open
Abstract
In cellulo crystallization is a rare event in nature. Recent advances that have made use of heterologous overexpression can promote the intracellular formation of protein crystals, but new tools are required to detect and characterize these targets in the complex cell environment. The present work makes use of Mask R-CNN, a convolutional neural network (CNN)-based instance segmentation method, for the identification of either single or multi-shaped crystals growing in living insect cells, using conventional bright field images. The algorithm can be rapidly adapted to recognize different targets, with the aim of extracting relevant information to support a semi-automated screening pipeline, in order to aid the development of the intracellular protein crystallization approach.
Collapse
Affiliation(s)
- Amirhossein Kardoost
- Sample Environment and Characterization Group, European XFEL GmbH, Holzkoppel 4, 22869 Schenefeld, Schleswig-Holstein, Germany
| | - Robert Schönherr
- Institute of Biochemistry, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Schleswig-Holstein, Germany
| | - Carsten Deiter
- Sample Environment and Characterization Group, European XFEL GmbH, Holzkoppel 4, 22869 Schenefeld, Schleswig-Holstein, Germany
| | - Lars Redecke
- Institute of Biochemistry, University of Lübeck, Ratzeburger Allee 160, 23562 Lübeck, Schleswig-Holstein, Germany
- Deutsches Elektronen-Synchrotron DESY, Photon Science, Notkestrasse 85, 22607 Hamburg, Germany
| | - Kristina Lorenzen
- Sample Environment and Characterization Group, European XFEL GmbH, Holzkoppel 4, 22869 Schenefeld, Schleswig-Holstein, Germany
| | - Joachim Schulz
- Sample Environment and Characterization Group, European XFEL GmbH, Holzkoppel 4, 22869 Schenefeld, Schleswig-Holstein, Germany
| | - Iñaki de Diego
- Sample Environment and Characterization Group, European XFEL GmbH, Holzkoppel 4, 22869 Schenefeld, Schleswig-Holstein, Germany
| |
Collapse
|
4
|
Alhakbani N, Alghamdi M, Al-Nafjan A. Design and Development of an Imitation Detection System for Human Action Recognition Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2023; 23:9889. [PMID: 38139734 PMCID: PMC10747182 DOI: 10.3390/s23249889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Revised: 11/22/2023] [Accepted: 12/14/2023] [Indexed: 12/24/2023]
Abstract
Human action recognition (HAR) is a rapidly growing field with numerous applications in various domains. HAR involves the development of algorithms and techniques to automatically identify and classify human actions from video data. Accurate recognition of human actions has significant implications in fields such as surveillance and sports analysis and in the health care domain. This paper presents a study on the design and development of an imitation detection system using an HAR algorithm based on deep learning. This study explores the use of deep learning models, such as a single-frame convolutional neural network (CNN) and pretrained VGG-16, for the accurate classification of human actions. The proposed models were evaluated using a benchmark dataset, KTH. The performance of these models was compared with that of classical classifiers, including K-Nearest Neighbors, Support Vector Machine, and Random Forest. The results showed that the VGG-16 model achieved higher accuracy than the single-frame CNN, with a 98% accuracy rate.
Collapse
Affiliation(s)
- Noura Alhakbani
- Information Technology Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia; (N.A.); (M.A.)
| | - Maha Alghamdi
- Information Technology Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia; (N.A.); (M.A.)
| | - Abeer Al-Nafjan
- Computer Science Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University, Riyadh 11432, Saudi Arabia
| |
Collapse
|
5
|
Balasubramaniam S, Velmurugan Y, Jaganathan D, Dhanasekaran S. A Modified LeNet CNN for Breast Cancer Diagnosis in Ultrasound Images. Diagnostics (Basel) 2023; 13:2746. [PMID: 37685284 PMCID: PMC10486538 DOI: 10.3390/diagnostics13172746] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Revised: 07/06/2023] [Accepted: 07/11/2023] [Indexed: 09/10/2023] Open
Abstract
Convolutional neural networks (CNNs) have been extensively utilized in medical image processing to automatically extract meaningful features and classify various medical conditions, enabling faster and more accurate diagnoses. In this paper, LeNet, a classic CNN architecture, has been successfully applied to breast cancer data analysis. It demonstrates its ability to extract discriminative features and classify malignant and benign tumors with high accuracy, thereby supporting early detection and diagnosis of breast cancer. LeNet with corrected Rectified Linear Unit (ReLU), a modification of the traditional ReLU activation function, has been found to improve the performance of LeNet in breast cancer data analysis tasks via addressing the "dying ReLU" problem and enhancing the discriminative power of the extracted features. This has led to more accurate, reliable breast cancer detection and diagnosis and improved patient outcomes. Batch normalization improves the performance and training stability of small and shallow CNN architecture like LeNet. It helps to mitigate the effects of internal covariate shift, which refers to the change in the distribution of network activations during training. This classifier will lessen the overfitting problem and reduce the running time. The designed classifier is evaluated against the benchmarking deep learning models, proving that this has produced a higher recognition rate. The accuracy of the breast image recognition rate is 89.91%. This model will achieve better performance in segmentation, feature extraction, classification, and breast cancer tumor detection.
Collapse
Affiliation(s)
| | - Yuvarajan Velmurugan
- Computer Science and Engineering, Sona College of Technology, Salem 636005, India; (Y.V.); (D.J.)
| | - Dhayanithi Jaganathan
- Computer Science and Engineering, Sona College of Technology, Salem 636005, India; (Y.V.); (D.J.)
| | | |
Collapse
|
6
|
Khan A, Chefranov A, Demirel H. Building discriminative features of scene recognition using multi-stages of inception-ResNet-v2. APPL INTELL 2023. [DOI: 10.1007/s10489-023-04460-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/31/2023]
|
7
|
Shehzad F, Attique Khan M, Asfand E. Yar M, Sharif M, Alhaisoni M, Tariq U, Majumdar A, Thinnukool O. Two-Stream Deep Learning Architecture-Based Human Action Recognition. COMPUTERS, MATERIALS & CONTINUA 2023; 74:5931-5949. [DOI: 10.32604/cmc.2023.028743] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Accepted: 05/06/2022] [Indexed: 08/25/2024]
|
8
|
Ahsan T, Khalid S, Najam S, Attique Khan M, Jin Kim Y, Chang B. HRNetO: Human Action Recognition Using Unified Deep Features Optimization Framework. COMPUTERS, MATERIALS & CONTINUA 2023; 75:1089-1105. [DOI: 10.32604/cmc.2023.034563] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Accepted: 12/08/2022] [Indexed: 08/25/2024]
|
9
|
Zhang J, Fang Z, Wang Z. Multi-feature fusion enhanced transformer with multi-layer fused decoding for image captioning. APPL INTELL 2022. [DOI: 10.1007/s10489-022-04202-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
10
|
Zhou D, Chen G, Xu F. Application of Deep Learning Technology in Strength Training of Football Players and Field Line Detection of Football Robots. Front Neurorobot 2022; 16:867028. [PMID: 35845757 PMCID: PMC9278879 DOI: 10.3389/fnbot.2022.867028] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2022] [Accepted: 04/19/2022] [Indexed: 11/25/2022] Open
Abstract
The purpose of the study is to improve the performance of intelligent football training. Based on deep learning (DL), the training of football players and detection of football robots are analyzed. First, the research status of the training of football players and football robots is introduced, and the basic structure of the neuron model and convolutional neural network (CNN) and the mainstream framework of DL are mainly expounded. Second, combined with the spatial stream network, a CNN-based action recognition system is constructed in the context of artificial intelligence (AI). Finally, by the football robot, a field line detection model based on a fully convolutional network (FCN) is proposed, and the effective applicability of the system is evaluated. The results demonstrate that the recognition effect of the dual-stream network is the best, reaching 92.8%. The recognition rate of the timestream network is lower than that of the dual-stream network, and the maximum recognition rate is 88%. The spatial stream network has the lowest recognition rate of 86.5%. The processing power of the four different algorithms on the dataset is stronger than that of the ordinary video set. The recognition rate of the time-segmented dual-stream fusion network is the highest, which is second only to the designed network. The recognition rate of the basic dual-stream network is 88.6%, and the recognition rate of the 3D CNN is the lowest, which is 86.2%. Under the intelligent training system, the recognition accuracy rates of jumping, kicking, grabbing, and starting actions range to 97.6, 94.5, 92.5, and 89.8% respectively, which are slightly lower than other actions. The recognition accuracy rate of passing action is 91.3%, and the maximum upgrade rate of intelligent training is 25.7%. The pixel accuracy of the improved field line detection of the model and the mean intersection over union (MIoU) are both improved by 5%. Intelligent training systems and the field line detection of football robots are more feasible. The research provides a reference for the development of AI in the field of sports training.
Collapse
Affiliation(s)
- Daliang Zhou
- School of PE, Nanjing Xiaozhuang University, Nanjing, China
| | - Gang Chen
- School of PE, Nanjing Xiaozhuang University, Nanjing, China
| | - Fei Xu
- School of Physical Education, Hangzhou Normal University, Hangzhou, China
| |
Collapse
|
11
|
Krzeszowski T, Switonski A, Kepski M, Calafate CT. Intelligent Sensors for Human Motion Analysis. SENSORS (BASEL, SWITZERLAND) 2022; 22:4952. [PMID: 35808444 PMCID: PMC9269847 DOI: 10.3390/s22134952] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/24/2022] [Accepted: 06/27/2022] [Indexed: 06/15/2023]
Abstract
Currently, the analysis of human motion is one of the most interesting and active research topics in computer science, especially in computer vision [...].
Collapse
Affiliation(s)
- Tomasz Krzeszowski
- Faculty of Electrical and Computer Engineering, Rzeszow University of Technology, 35-959 Rzeszow, Poland
| | - Adam Switonski
- Department of Graphics, Computer Vision and Digital Systems, Silesian University of Technology, 44-100 Gliwice, Poland;
| | - Michal Kepski
- Institute of Computer Science, University of Rzeszów, 35-310 Rzeszow, Poland;
| | - Carlos T. Calafate
- Computer Engineering Department, Universitat Politècnica de València (UPV), 46022 Valencia, Spain;
| |
Collapse
|
12
|
Amraee S, Chinipardaz M, Charoosaei M. Analytical study of two feature extraction methods in comparison with deep learning methods for classification of small metal objects. Vis Comput Ind Biomed Art 2022; 5:13. [PMID: 35534747 PMCID: PMC9085991 DOI: 10.1186/s42492-022-00111-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2021] [Accepted: 04/26/2022] [Indexed: 11/20/2022] Open
Abstract
This paper addresses the efficiency of two feature extraction methods for classifying small metal objects including screws, nuts, keys, and coins: the histogram of oriented gradients (HOG) and local binary pattern (LBP). The desired features for the labeled images are first extracted and saved in the form of a feature matrix. Using three different classification methods (non-parametric K-nearest neighbors algorithm, support vector machine, and naïve Bayesian method), the images are classified into four different classes. Then, by examining the resulting confusion matrix, the performances of the HOG and LBP approaches are compared for these four classes. The effectiveness of these two methods is also compared with the “You Only Look Once” and faster region-based convolutional neural network approaches, which are based on deep learning. The collected image set in this paper includes 800 labeled training images and 180 test images. The results show that the use of the HOG is more efficient than the use of the LBP. Moreover, a combination of the HOG and LBP provides better results than either alone.
Collapse
Affiliation(s)
- Somaieh Amraee
- Department of Electrical and Computer Engineering, Jundi-Shapur University of Technology, Dezful, 64615/334, Iran.
| | - Maryam Chinipardaz
- Department of Electrical and Computer Engineering, Jundi-Shapur University of Technology, Dezful, 64615/334, Iran
| | - Mohammadali Charoosaei
- Department of Electrical and Computer Engineering, Jundi-Shapur University of Technology, Dezful, 64615/334, Iran
| |
Collapse
|
13
|
Nasir IM, Raza M, Shah JH, Wang SH, Tariq U, Khan MA. HAREDNet: A deep learning based architecture for autonomous video surveillance by recognizing human actions. COMPUTERS AND ELECTRICAL ENGINEERING 2022; 99:107805. [DOI: 10.1016/j.compeleceng.2022.107805] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/25/2024]
|
14
|
KhoKhar FA, Shah JH, Khan MA, Sharif M, Tariq U, Kadry S. A review on federated learning towards image processing. COMPUTERS AND ELECTRICAL ENGINEERING 2022; 99:107818. [DOI: 10.1016/j.compeleceng.2022.107818] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/25/2024]
|
15
|
Anthropometric Ratios for Lower-Body Detection Based on Deep Learning and Traditional Methods. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12052678] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Lower-body detection can be useful in many applications, such as the detection of falling and injuries during exercises. However, it can be challenging to detect the lower-body, especially under various lighting and occlusion conditions. This paper presents a novel lower-body detection framework using proposed anthropometric ratios and compares the performance of deep learning (convolutional neural networks and OpenPose) and traditional detection methods. According to the results, the proposed framework helps to successfully detect the accurate boundaries of the lower-body under various illumination and occlusion conditions for lower-limb monitoring. The proposed framework of anthropometric ratios combined with convolutional neural networks (A-CNNs) also achieves high accuracy (90.14%), while the combination of anthropometric ratios and traditional techniques (A-Traditional) for lower-body detection shows satisfactory performance with an averaged accuracy (74.81%). Although the accuracy of OpenPose (95.82%) is higher than the A-CNNs for lower-body detection, the A-CNNs provides lower complexity than the OpenPose, which is advantageous for lower-body detection and implementation on monitoring systems.
Collapse
|
16
|
Abstract
Human Activity Recognition (HAR) is the process of identifying human actions in a specific environment. Recognizing human activities from video streams is a challenging task due to problems such as background noise, partial occlusion, changes in scale, orientation, lighting, and the unstable capturing process. Such multi-dimensional and none-linear process increases the complexity, making traditional solutions inefficient in terms of several performance indicators such as accuracy, time, and memory. This paper proposes a technique to select a set of representative features that can accurately recognize human activities from video streams, while minimizing the recognition time and memory. The extracted features are projected on a canvas, which keeps the synchronization property of the spatiotemporal information. The proposed technique is developed to select the features that refer only to progression of changes. The original RGB frames are preprocessed using background subtraction to extract the subject. Then the activity pattern is extracted through the proposed Growth method. Three experiments were conducted; the first experiment was a baseline to compare the classification task using the original RGB features. The second experiment relied on classifying activities using the proposed feature-selection method. Finally, the third experiment provided a sensitivity analysis that compares between the effect of both techniques on time and memory resources. The results indicated that the proposed method outperformed original RBG feature-selection method in terms of accuracy, time, and memory requirements.
Collapse
|
17
|
Image Retrieval via Canonical Correlation Analysis and Binary Hypothesis Testing. INFORMATION 2022. [DOI: 10.3390/info13030106] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Abstract
Canonical Correlation Analysis (CCA) is a classic multivariate statistical technique, which can be used to find a projection pair that maximally captures the correlation between two sets of random variables. The present paper introduces a CCA-based approach for image retrieval. It capitalizes on feature maps induced by two images under comparison through a pre-trained Convolutional Neural Network (CNN) and leverages basis vectors identified through CCA, together with an element-wise selection method based on a Chernoff-information-related criterion, to produce compact transformed image features; a binary hypothesis test regarding the joint distribution of transformed feature pair is then employed to measure the similarity between two images. The proposed approach is benchmarked against two alternative statistical methods, Linear Discriminant Analysis (LDA) and Principal Component Analysis with whitening (PCAw). Our CCA-based approach is shown to achieve highly competitive retrieval performances on standard datasets, which include, among others, Oxford5k and Paris6k.
Collapse
|
18
|
Breast Cancer Mammograms Classification Using Deep Neural Network and Entropy-Controlled Whale Optimization Algorithm. Diagnostics (Basel) 2022; 12:diagnostics12020557. [PMID: 35204646 PMCID: PMC8871265 DOI: 10.3390/diagnostics12020557] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2021] [Revised: 01/22/2022] [Accepted: 01/30/2022] [Indexed: 02/04/2023] Open
Abstract
Breast cancer has affected many women worldwide. To perform detection and classification of breast cancer many computer-aided diagnosis (CAD) systems have been established because the inspection of the mammogram images by the radiologist is a difficult and time taken task. To early diagnose the disease and provide better treatment lot of CAD systems were established. There is still a need to improve existing CAD systems by incorporating new methods and technologies in order to provide more precise results. This paper aims to investigate ways to prevent the disease as well as to provide new methods of classification in order to reduce the risk of breast cancer in women's lives. The best feature optimization is performed to classify the results accurately. The CAD system's accuracy improved by reducing the false-positive rates.The Modified Entropy Whale Optimization Algorithm (MEWOA) is proposed based on fusion for deep feature extraction and perform the classification. In the proposed method, the fine-tuned MobilenetV2 and Nasnet Mobile are applied for simulation. The features are extracted, and optimization is performed. The optimized features are fused and optimized by using MEWOA. Finally, by using the optimized deep features, the machine learning classifiers are applied to classify the breast cancer images. To extract the features and perform the classification, three publicly available datasets are used: INbreast, MIAS, and CBIS-DDSM. The maximum accuracy achieved in INbreast dataset is 99.7%, MIAS dataset has 99.8% and CBIS-DDSM has 93.8%. Finally, a comparison with other existing methods is performed, demonstrating that the proposed algorithm outperforms the other approaches.
Collapse
|
19
|
A Review of Deep Learning Methods for Compressed Sensing Image Reconstruction and Its Medical Applications. ELECTRONICS 2022. [DOI: 10.3390/electronics11040586] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Compressed sensing (CS) and its medical applications are active areas of research. In this paper, we review recent works using deep learning method to solve CS problem for images or medical imaging reconstruction including computed tomography (CT), magnetic resonance imaging (MRI) and positron-emission tomography (PET). We propose a novel framework to unify traditional iterative algorithms and deep learning approaches. In short, we define two projection operators toward image prior and data consistency, respectively, and any reconstruction algorithm can be decomposed to the two parts. Though deep learning methods can be divided into several categories, they all satisfies the framework. We built the relationship between different reconstruction methods of deep learning, and connect them to traditional methods through the proposed framework. It also indicates that the key to solve CS problem and its medical applications is how to depict the image prior. Based on the framework, we analyze the current deep learning methods and point out some important directions of research in the future.
Collapse
|
20
|
Masood H, Zafar A, Ali MU, Hussain T, Khan MA, Tariq U, Damaševičius R. Tracking of a Fixed-Shape Moving Object Based on the Gradient Descent Method. SENSORS (BASEL, SWITZERLAND) 2022; 22:1098. [PMID: 35161843 PMCID: PMC8839945 DOI: 10.3390/s22031098] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/30/2021] [Revised: 01/25/2022] [Accepted: 01/28/2022] [Indexed: 11/19/2022]
Abstract
Tracking moving objects is one of the most promising yet the most challenging research areas pertaining to computer vision, pattern recognition and image processing. The challenges associated with object tracking range from problems pertaining to camera axis orientations to object occlusion. In addition, variations in remote scene environments add to the difficulties related to object tracking. All the mentioned challenges and problems pertaining to object tracking make the procedure computationally complex and time-consuming. In this paper, a stochastic gradient-based optimization technique has been used in conjunction with particle filters for object tracking. First, the object that needs to be tracked is detected using the Maximum Average Correlation Height (MACH) filter. The object of interest is detected based on the presence of a correlation peak and average similarity measure. The results of object detection are fed to the tracking routine. The gradient descent technique is employed for object tracking and is used to optimize the particle filters. The gradient descent technique allows particles to converge quickly, allowing less time for the object to be tracked. The results of the proposed algorithm are compared with similar state-of-the-art tracking algorithms on five datasets that include both artificial moving objects and humans to show that the gradient-based tracking algorithm provides better results, both in terms of accuracy and speed.
Collapse
Affiliation(s)
- Haris Masood
- Electrical Engineering Department, Wah Engineering College, University of Wah, Wah Cantt 47040, Pakistan; (H.M.); (T.H.)
| | - Amad Zafar
- Department of Electrical Engineering, The Ibadat International University, Islamabad 54590, Pakistan;
| | - Muhammad Umair Ali
- Department of Unmanned Vehicle Engineering, Sejong University, Seoul 05006, Korea;
| | - Tehseen Hussain
- Electrical Engineering Department, Wah Engineering College, University of Wah, Wah Cantt 47040, Pakistan; (H.M.); (T.H.)
| | | | - Usman Tariq
- College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia;
| | - Robertas Damaševičius
- Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania
| |
Collapse
|
21
|
Deep Learning and Kurtosis-Controlled, Entropy-Based Framework for Human Gait Recognition Using Video Sequences. ELECTRONICS 2022. [DOI: 10.3390/electronics11030334] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Gait is commonly defined as the movement pattern of the limbs over a hard substrate, and it serves as a source of identification information for various computer-vision and image-understanding techniques. A variety of parameters, such as human clothing, angle shift, walking style, occlusion, and so on, have a significant impact on gait-recognition systems, making the scene quite complex to handle. In this article, we propose a system that effectively handles problems associated with viewing angle shifts and walking styles in a real-time environment. The following steps are included in the proposed novel framework: (a) real-time video capture, (b) feature extraction using transfer learning on the ResNet101 deep model, and (c) feature selection using the proposed kurtosis-controlled entropy (KcE) approach, followed by a correlation-based feature fusion step. The most discriminant features are then classified using the most advanced machine learning classifiers. The simulation process is fed by the CASIA B dataset as well as a real-time captured dataset. On selected datasets, the accuracy is 95.26% and 96.60%, respectively. When compared to several known techniques, the results show that our proposed framework outperforms them all.
Collapse
|
22
|
Cucumber Leaf Diseases Recognition Using Multi Level Deep Entropy-ELM Feature Selection. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12020593] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Agriculture has becomes an immense area of research and is ascertained as a key element in the area of computer vision. In the agriculture field, image processing acts as a primary part. Cucumber is an important vegetable and its production in Pakistan is higher as compared to the other vegetables because of its use in salads. However, the diseases of cucumber such as Angular leaf spot, Anthracnose, blight, Downy mildew, and powdery mildew widely decrease the quality and quantity. Lately, numerous methods have been proposed for the identification and classification of diseases. Early detection and then treatment of the diseases in plants is important to prevent the crop from a disastrous decrease in yields. Many classification techniques have been proposed but still, they are facing some challenges such as noise, redundant features, and extraction of relevant features. In this work, an automated framework is proposed using deep learning and best feature selection for cucumber leaf diseases classification. In the proposed framework, initially, an augmentation technique is applied to the original images by creating more training data from existing samples and handling the problem of the imbalanced dataset. Then two different phases are utilized. In the first phase, fine-tuned four pre-trained models and select the best of them based on the accuracy. Features are extracted from the selected fine-tuned model and refined through the Entropy-ELM technique. In the second phase, fused the features of all four fine-tuned models and apply the Entropy-ELM technique, and finally fused with phase 1 selected feature. Finally, the fused features are recognized using machine learning classifiers for the final classification. The experimental process is conducted on five different datasets. On these datasets, the best-achieved accuracy is 98.4%. The proposed framework is evaluated on each step and also compared with some recent techniques. The comparison with some recent techniques showed that the proposed method obtained an improved performance.
Collapse
|
23
|
Naeem Akbar M, Riaz F, Bilal Awan A, Attique Khan M, Tariq U, Rehman S. A Hybrid Duo-Deep Learning and Best Features Based Framework for燗ction燫ecognition. COMPUTERS, MATERIALS & CONTINUA 2022; 73:2555-2576. [DOI: 10.32604/cmc.2022.028696] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/25/2024]
|