1
|
Almujally NA, Rafique AA, Al Mudawi N, Alazeb A, Alonazi M, Algarni A, Jalal A, Liu H. Multi-modal remote perception learning for object sensory data. Front Neurorobot 2024; 18:1427786. [PMID: 39377028 PMCID: PMC11457376 DOI: 10.3389/fnbot.2024.1427786] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Accepted: 08/26/2024] [Indexed: 10/09/2024] Open
Abstract
Introduction When it comes to interpreting visual input, intelligent systems make use of contextual scene learning, which significantly improves both resilience and context awareness. The management of enormous amounts of data is a driving force behind the growing interest in computational frameworks, particularly in the context of autonomous cars. Method The purpose of this study is to introduce a novel approach known as Deep Fused Networks (DFN), which improves contextual scene comprehension by merging multi-object detection and semantic analysis. Results To enhance accuracy and comprehension in complex situations, DFN makes use of a combination of deep learning and fusion techniques. With a minimum gain of 6.4% in accuracy for the SUN-RGB-D dataset and 3.6% for the NYU-Dv2 dataset. Discussion Findings demonstrate considerable enhancements in object detection and semantic analysis when compared to the methodologies that are currently being utilized.
Collapse
Affiliation(s)
- Nouf Abdullah Almujally
- Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Adnan Ahmed Rafique
- Department of Computer Science and IT, University of Poonch Rawalakot, Rawalakot, Pakistan
| | - Naif Al Mudawi
- Department of Computer Science, College of Computer Science and Information System, Najran University, Najran, Saudi Arabia
| | - Abdulwahab Alazeb
- Department of Computer Science, College of Computer Science and Information System, Najran University, Najran, Saudi Arabia
| | - Mohammed Alonazi
- Department of Computer Engineering, College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | - Asaad Algarni
- Department of Computer Sciences, Faculty of Computing and Information Technology, Northern Border University, Rafha, Saudi Arabia
| | - Ahmad Jalal
- Faculty of Computer Science, Air University, Islamabad, Pakistan
- Department of Computer Science and Engineering, College of Informatics, Korea University, Seoul, Republic of Korea
| | - Hui Liu
- Cognitive Systems Lab, University of Bremen, Bremen, Germany
| |
Collapse
|
2
|
Su Z, Adam A, Nasrudin MF, Prabuwono AS. Proposal-Free Fully Convolutional Network: Object Detection Based on a Box Map. SENSORS (BASEL, SWITZERLAND) 2024; 24:3529. [PMID: 38894319 PMCID: PMC11175249 DOI: 10.3390/s24113529] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/13/2024] [Revised: 05/23/2024] [Accepted: 05/23/2024] [Indexed: 06/21/2024]
Abstract
Region proposal-based detectors, such as Region-Convolutional Neural Networks (R-CNNs), Fast R-CNNs, Faster R-CNNs, and Region-Based Fully Convolutional Networks (R-FCNs), employ a two-stage process involving region proposal generation followed by classification. This approach is effective but computationally intensive and typically slower than proposal-free methods. Therefore, region proposal-free detectors are becoming popular to balance accuracy and speed. This paper proposes a proposal-free, fully convolutional network (PF-FCN) that outperforms other state-of-the-art, proposal-free methods. Unlike traditional region proposal-free methods, PF-FCN can generate a "box map" based on regression training techniques. This box map comprises a set of vectors, each designed to produce bounding boxes corresponding to the positions of objects in the input image. The channel and spatial contextualized sub-network are further designed to learn a "box map". In comparison to renowned proposal-free detectors such as CornerNet, CenterNet, and You Look Only Once (YOLO), PF-FCN utilizes a fully convolutional, single-pass method. By reducing the need for fully connected layers and filtering center points, the method considerably reduces the number of trained parameters and optimizes the scalability across varying input sizes. Evaluations of benchmark datasets suggest the effectiveness of PF-FCN: the proposed model achieved an mAP of 89.6% on PASCAL VOC 2012 and 71.7% on MS COCO, which are higher than those of the baseline Fully Convolutional One-Stage Detector (FCOS) and other classical proposal-free detectors. The results prove the significance of proposal-free detectors in both practical applications and future research.
Collapse
Affiliation(s)
- Zhihao Su
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.)
| | - Afzan Adam
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.)
| | - Mohammad Faidzul Nasrudin
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.)
| | - Anton Satria Prabuwono
- Faculty of Computing and Information Technology in Rabigh, King Abdulaziz University, Rabigh 21911, Saudi Arabia;
| |
Collapse
|
3
|
Tel A, Raccampo L, Vinayahalingam S, Troise S, Abbate V, Orabona GD, Sembronio S, Robiony M. Complex Craniofacial Cases through Augmented Reality Guidance in Surgical Oncology: A Technical Report. Diagnostics (Basel) 2024; 14:1108. [PMID: 38893634 PMCID: PMC11171943 DOI: 10.3390/diagnostics14111108] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 05/20/2024] [Accepted: 05/24/2024] [Indexed: 06/21/2024] Open
Abstract
Augmented reality (AR) is a promising technology to enhance image guided surgery and represents the perfect bridge to combine precise virtual planning with computer-aided execution of surgical maneuvers in the operating room. In craniofacial surgical oncology, AR brings to the surgeon's sight a digital, three-dimensional representation of the anatomy and helps to identify tumor boundaries and optimal surgical paths. Intraoperatively, real-time AR guidance provides surgeons with accurate spatial information, ensuring accurate tumor resection and preservation of critical structures. In this paper, the authors review current evidence of AR applications in craniofacial surgery, focusing on real surgical applications, and compare existing literature with their experience during an AR and navigation guided craniofacial resection, to subsequently analyze which technological trajectories will represent the future of AR and define new perspectives of application for this revolutionizing technology.
Collapse
Affiliation(s)
- Alessandro Tel
- Clinic of Maxillofacial Surgery, Head-Neck and NeuroScience Department, University Hospital of Udine, p.le S. Maria della Misericordia 15, 33100 Udine, Italy
| | - Luca Raccampo
- Clinic of Maxillofacial Surgery, Head-Neck and NeuroScience Department, University Hospital of Udine, p.le S. Maria della Misericordia 15, 33100 Udine, Italy
| | - Shankeeth Vinayahalingam
- Department of Oral and Maxillofacial Surgery, Radboud University Medical Center, 6525 GA Nijmegen, The Netherlands
| | - Stefania Troise
- Neurosciences Reproductive and Odontostomatological Sciences Department, University of Naples “Federico II”, 80131 Naples, Italy
| | - Vincenzo Abbate
- Neurosciences Reproductive and Odontostomatological Sciences Department, University of Naples “Federico II”, 80131 Naples, Italy
| | - Giovanni Dell’Aversana Orabona
- Neurosciences Reproductive and Odontostomatological Sciences Department, University of Naples “Federico II”, 80131 Naples, Italy
| | - Salvatore Sembronio
- Clinic of Maxillofacial Surgery, Head-Neck and NeuroScience Department, University Hospital of Udine, p.le S. Maria della Misericordia 15, 33100 Udine, Italy
| | - Massimo Robiony
- Clinic of Maxillofacial Surgery, Head-Neck and NeuroScience Department, University Hospital of Udine, p.le S. Maria della Misericordia 15, 33100 Udine, Italy
| |
Collapse
|
4
|
Naich AY, Carrión JR. LiDAR-Based Intensity-Aware Outdoor 3D Object Detection. SENSORS (BASEL, SWITZERLAND) 2024; 24:2942. [PMID: 38733047 PMCID: PMC11086319 DOI: 10.3390/s24092942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/05/2024] [Revised: 04/28/2024] [Accepted: 05/01/2024] [Indexed: 05/13/2024]
Abstract
LiDAR-based 3D object detection and localization are crucial components of autonomous navigation systems, including autonomous vehicles and mobile robots. Most existing LiDAR-based 3D object detection and localization approaches primarily use geometric or structural feature abstractions from LiDAR point clouds. However, these approaches can be susceptible to environmental noise due to adverse weather conditions or the presence of highly scattering media. In this work, we propose an intensity-aware voxel encoder for robust 3D object detection. The proposed voxel encoder generates an intensity histogram that describes the distribution of point intensities within a voxel and is used to enhance the voxel feature set. We integrate this intensity-aware encoder into an efficient single-stage voxel-based detector for 3D object detection. Experimental results obtained using the KITTI dataset show that our method achieves comparable results with respect to the state-of-the-art method for car objects in 3D detection and from a bird's-eye view and superior results for pedestrian and cyclic objects. Furthermore, our model can achieve a detection rate of 40.7 FPS during inference time, which is higher than that of the state-of-the-art methods and incurs a lower computational cost.
Collapse
Affiliation(s)
- Ammar Yasir Naich
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK
| | - Jesús Requena Carrión
- School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK
| |
Collapse
|
5
|
Harfouche AL, Nakhle F, Corona P. Metaverse technology innovating plant science research and learning. TRENDS IN PLANT SCIENCE 2024; 29:266-267. [PMID: 37821337 DOI: 10.1016/j.tplants.2023.09.009] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 09/05/2023] [Accepted: 09/08/2023] [Indexed: 10/13/2023]
Affiliation(s)
- Antoine L Harfouche
- Department for Innovation in Biological, Agro-food, and Forest systems, University of Tuscia, Viterbo 01100, Italy.
| | - Farid Nakhle
- Department for Innovation in Biological, Agro-food, and Forest systems, University of Tuscia, Viterbo 01100, Italy
| | - Piermaria Corona
- Research Centre for Forestry and Wood, Council for Agricultural Research and Economics, I-52100 Arezzo, Italy
| |
Collapse
|
6
|
Malta A, Farinha T, Mendes M. Augmented Reality in Maintenance-History and Perspectives. J Imaging 2023; 9:142. [PMID: 37504819 PMCID: PMC10381749 DOI: 10.3390/jimaging9070142] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Revised: 06/23/2023] [Accepted: 07/03/2023] [Indexed: 07/29/2023] Open
Abstract
Augmented Reality (AR) is a technology that allows virtual elements to be superimposed over images of real contexts, whether these are text elements, graphics, or other types of objects. Smart AR glasses are increasingly optimized, and modern ones have features such as Global Positioning System (GPS), a microphone, and gesture recognition, among others. These devices allow users to have their hands free to perform tasks while they receive instructions in real time through the glasses. This allows maintenance professionals to carry out interventions more efficiently and in a shorter time than would be necessary without the support of this technology. In the present work, a timeline of important achievements is established, including important findings in object recognition, real-time operation. and integration of technologies for shop floor use. Perspectives on future research and related recommendations are proposed as well.
Collapse
Affiliation(s)
- Ana Malta
- Coimbra Institute of Engineering, Rua Pedro Nunes-Quinta da Nora, Polytechnic Institute of Coimbra, 3030-199 Coimbra, Portugal
- RCM2+ Research Centre for Asset Management and Systems Engineering, ISEC/IPC, Rua Pedro Nunes, 3030-199 Coimbra, Portugal
| | - Torres Farinha
- Coimbra Institute of Engineering, Rua Pedro Nunes-Quinta da Nora, Polytechnic Institute of Coimbra, 3030-199 Coimbra, Portugal
- RCM2+ Research Centre for Asset Management and Systems Engineering, ISEC/IPC, Rua Pedro Nunes, 3030-199 Coimbra, Portugal
| | - Mateus Mendes
- Coimbra Institute of Engineering, Rua Pedro Nunes-Quinta da Nora, Polytechnic Institute of Coimbra, 3030-199 Coimbra, Portugal
- RCM2+ Research Centre for Asset Management and Systems Engineering, ISEC/IPC, Rua Pedro Nunes, 3030-199 Coimbra, Portugal
| |
Collapse
|
7
|
Seetohul J, Shafiee M, Sirlantzis K. Augmented Reality (AR) for Surgical Robotic and Autonomous Systems: State of the Art, Challenges, and Solutions. SENSORS (BASEL, SWITZERLAND) 2023; 23:6202. [PMID: 37448050 DOI: 10.3390/s23136202] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/22/2023] [Revised: 06/09/2023] [Accepted: 07/03/2023] [Indexed: 07/15/2023]
Abstract
Despite the substantial progress achieved in the development and integration of augmented reality (AR) in surgical robotic and autonomous systems (RAS), the center of focus in most devices remains on improving end-effector dexterity and precision, as well as improved access to minimally invasive surgeries. This paper aims to provide a systematic review of different types of state-of-the-art surgical robotic platforms while identifying areas for technological improvement. We associate specific control features, such as haptic feedback, sensory stimuli, and human-robot collaboration, with AR technology to perform complex surgical interventions for increased user perception of the augmented world. Current researchers in the field have, for long, faced innumerable issues with low accuracy in tool placement around complex trajectories, pose estimation, and difficulty in depth perception during two-dimensional medical imaging. A number of robots described in this review, such as Novarad and SpineAssist, are analyzed in terms of their hardware features, computer vision systems (such as deep learning algorithms), and the clinical relevance of the literature. We attempt to outline the shortcomings in current optimization algorithms for surgical robots (such as YOLO and LTSM) whilst providing mitigating solutions to internal tool-to-organ collision detection and image reconstruction. The accuracy of results in robot end-effector collisions and reduced occlusion remain promising within the scope of our research, validating the propositions made for the surgical clearance of ever-expanding AR technology in the future.
Collapse
Affiliation(s)
- Jenna Seetohul
- Mechanical Engineering Group, School of Engineering, University of Kent, Canterbury CT2 7NT, UK
| | - Mahmood Shafiee
- Mechanical Engineering Group, School of Engineering, University of Kent, Canterbury CT2 7NT, UK
- School of Mechanical Engineering Sciences, University of Surrey, Guildford GU2 7XH, UK
| | - Konstantinos Sirlantzis
- School of Engineering, Technology and Design, Canterbury Christ Church University, Canterbury CT1 1QU, UK
- Intelligent Interactions Group, School of Engineering, University of Kent, Canterbury CT2 7NT, UK
| |
Collapse
|
8
|
Yao JF, Yang Y, Wang XC, Zhang XP. Systematic review of digital twin technology and applications. Vis Comput Ind Biomed Art 2023; 6:10. [PMID: 37249731 DOI: 10.1186/s42492-023-00137-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Accepted: 05/18/2023] [Indexed: 05/31/2023] Open
Abstract
As one of the most important applications of digitalization, intelligence, and service, the digital twin (DT) breaks through the constraints of time, space, cost, and security on physical entities, expands and optimizes the relevant functions of physical entities, and enhances their application value. This phenomenon has been widely studied in academia and industry. In this study, the concept and definition of DT, as utilized by scholars and researchers in various fields of industry, are summarized. The internal association between DT and related technologies is explained. The four stages of DT development history are identified. The fundamentals of the technology, evaluation indexes, and model frameworks are reviewed. Subsequently, a conceptual ternary model of DT based on time, space, and logic is proposed. The technology and application status of typical DT systems are described. Finally, the current technical challenges of DT technology are analyzed, and directions for future development are discussed.
Collapse
Affiliation(s)
- Jun-Feng Yao
- Center for Digital Media Computing, School of Film, Xiamen University, Xiamen 361005, China.
- School of Informatics, Xiamen University, Xiamen 361005, China.
- Key Laboratory of Digital Protection and Intelligent Processing of Intangible Cultural Heritage of Fujian and Taiwan, Ministry of Culture and Tourism, Xiamen 361005, China.
| | - Yong Yang
- Center for Digital Media Computing, School of Film, Xiamen University, Xiamen 361005, China
| | - Xue-Cheng Wang
- Center for Digital Media Computing, School of Film, Xiamen University, Xiamen 361005, China
| | - Xiao-Peng Zhang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, the Institute of Automation, Chinese Academy of Sciences, Beijing 101408, China
| |
Collapse
|
9
|
Li Q, Hu S, Shimasaki K, Ishii I. An Active Multi-Object Ultrafast Tracking System with CNN-Based Hybrid Object Detection. SENSORS (BASEL, SWITZERLAND) 2023; 23:4150. [PMID: 37112491 PMCID: PMC10145589 DOI: 10.3390/s23084150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Revised: 04/09/2023] [Accepted: 04/11/2023] [Indexed: 06/19/2023]
Abstract
This study proposes a visual tracking system that can detect and track multiple fast-moving appearance-varying targets simultaneously with 500 fps image processing. The system comprises a high-speed camera and a pan-tilt galvanometer system, which can rapidly generate large-scale high-definition images of the wide monitored area. We developed a CNN-based hybrid tracking algorithm that can robustly track multiple high-speed moving objects simultaneously. Experimental results demonstrate that our system can track up to three moving objects with velocities lower than 30 m per second simultaneously within an 8-m range. The effectiveness of our system was demonstrated through several experiments conducted on simultaneous zoom shooting of multiple moving objects (persons and bottles) in a natural outdoor scene. Moreover, our system demonstrates high robustness to target loss and crossing situations.
Collapse
Affiliation(s)
| | | | | | - Idaku Ishii
- Correspondence: ; Tel.: +81-82-424-7692; Fax: +81-82-422-7158
| |
Collapse
|
10
|
Xu X, Huo W, Li F, Zhou H. Classification of Liquid Ingress in GFRP Honeycomb Based on One-Dimension Sequential Model Using THz-TDS. SENSORS (BASEL, SWITZERLAND) 2023; 23:1149. [PMID: 36772188 PMCID: PMC9921085 DOI: 10.3390/s23031149] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/06/2022] [Revised: 01/12/2023] [Accepted: 01/18/2023] [Indexed: 06/18/2023]
Abstract
Honeycomb structure composites are taking an increasing proportion in aircraft manufacturing because of their high strength-to-weight ratio, good fatigue resistance, and low manufacturing cost. However, the hollow structure is very prone to liquid ingress. Here, we report a fast and automatic classification approach for water, alcohol, and oil filled in glass fiber reinforced polymer (GFRP) honeycomb structures through terahertz time-domain spectroscopy (THz-TDS). We propose an improved one-dimensional convolutional neural network (1D-CNN) model, and compared it with long short-term memory (LSTM) and ordinary 1D-CNN models, which are classification networks based on one dimension sequenced signals. The automated liquid classification results show that the LSTM model has the best performance for the time-domain signals, while the improved 1D-CNN model performed best for the frequency-domain signals.
Collapse
Affiliation(s)
- Xiaohui Xu
- School of Armament Science and Technology, Xi’an Technological University, Xi’an 710064, China
| | - Wenjun Huo
- School of Armament Science and Technology, Xi’an Technological University, Xi’an 710064, China
| | - Fei Li
- School of Mechatronic Engineering, Xi’an Technological University, Xi’an 710064, China
| | - Hongbin Zhou
- School of Equipment Management and UAV Engineering, Air Force Engineering University, Xi’an 710043, China
| |
Collapse
|
11
|
Xie J, Chai JJK, O’Sullivan C, Xu JL. Trends of Augmented Reality for Agri-Food Applications. SENSORS (BASEL, SWITZERLAND) 2022; 22:8333. [PMID: 36366030 PMCID: PMC9653656 DOI: 10.3390/s22218333] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/19/2022] [Revised: 10/25/2022] [Accepted: 10/27/2022] [Indexed: 06/16/2023]
Abstract
Recent years have witnessed an increasing interest in deploying state-of-the-art augmented reality (AR) head-mounted displays (HMDs) for agri-food applications. The benefits of AR HMDs to agri-food industry stakeholders (e.g., food suppliers, retail/food service) have received growing attention and recognition. AR HMDs enable users to make healthier dietary choices, experience novel changes in their perception of taste, enhance the cooking and food shopping experience, improve productivity at work and enhance the implementation of precision farming. Therefore, although development costs are still high, the case for integration of AR in food chains appears to be compelling. This review will present the most recent developments of AR HMDs for agri-food relevant applications. The summarized applications can be clustered into different themes: (1) dietary and food nutrition assessment; (2) food sensory science; (3) changing the eating environment; (4) retail food chain applications; (5) enhancing the cooking experience; (6) food-related training and learning; and (7) food production and precision farming. Limitations of current practices will be highlighted, along with some proposed applications.
Collapse
Affiliation(s)
- Junhao Xie
- School of Biosystems and Food Engineering, University College Dublin, Belfield, Dublin 4, Ireland
| | - Jackey J. K. Chai
- School of Computer Science and Statistics, Trinity College Dublin, College Green, Dublin 2, Ireland
| | - Carol O’Sullivan
- School of Computer Science and Statistics, Trinity College Dublin, College Green, Dublin 2, Ireland
| | - Jun-Li Xu
- School of Biosystems and Food Engineering, University College Dublin, Belfield, Dublin 4, Ireland
| |
Collapse
|
12
|
Lin P, Li D, Jia Y, Chen Y, Huang G, Elkhouchlaa H, Yao Z, Zhou Z, Zhou H, Li J, Lu H. A novel approach for estimating the flowering rate of litchi based on deep learning and UAV images. FRONTIERS IN PLANT SCIENCE 2022; 13:966639. [PMID: 36092399 PMCID: PMC9453484 DOI: 10.3389/fpls.2022.966639] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/11/2022] [Accepted: 07/29/2022] [Indexed: 06/15/2023]
Abstract
Litchi flowering management is an important link in litchi orchard management. Statistical litchi flowering rate data can provide an important reference for regulating the number of litchi flowers and directly determining the quality and yield of litchi fruit. At present, the statistical work regarding litchi flowering rates requires considerable labour costs. Therefore, this study aims at the statistical litchi flowering rate task, and a combination of unmanned aerial vehicle (UAV) images and computer vision technology is proposed to count the numbers of litchi flower clusters and flushes in a complex natural environment to improve the efficiency of litchi flowering rate estimation. First, RGB images of litchi canopies at the flowering stage are collected by a UAV. After performing image preprocessing, a dataset is established, and two types of objects in the images, namely, flower clusters and flushes, are manually labelled. Second, by comparing the pretraining and testing results obtained when setting different training parameters for the YOLOv4 model, the optimal parameter combination is determined. The YOLOv4 model trained with the optimal combination of parameters tests best on the test set, at which time the mean average precision (mAP) is 87.87%. The detection time required for a single image is 0.043 s. Finally, aiming at the two kinds of targets (flower clusters and flushes) on 8 litchi trees in a real orchard, a model for estimating the numbers of flower clusters and flushes on a single litchi tree is constructed by matching the identified number of targets with the actual number of targets via equation fitting. Then, the data obtained from the manual counting process and the estimation model for the other five litchi trees in the real orchard are statistically analysed. The average error rate for the number of flower clusters is 4.20%, the average error rate for the number of flushes is 2.85%, and the average error for the flowering rate is 1.135%. The experimental results show that the proposed method is effective for estimating the litchi flowering rate and can provide guidance regarding the management of the flowering periods of litchi orchards.
Collapse
Affiliation(s)
- Peiyi Lin
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Denghui Li
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Yuhang Jia
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Yingyi Chen
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Guangwen Huang
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Hamza Elkhouchlaa
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Zhongwei Yao
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Zhengqi Zhou
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Haobo Zhou
- College of Engineering, South China Agricultural University, Guangzhou, China
| | - Jun Li
- College of Engineering, South China Agricultural University, Guangzhou, China
- Guangdong Laboratory for Lingnan Modern Agriculture, Guangzhou, China
| | - Huazhong Lu
- Guangdong Academy of Agricultural Sciences, Guangzhou, China
| |
Collapse
|
13
|
Review of Artificial Intelligence and Machine Learning Technologies: Classification, Restrictions, Opportunities and Challenges. MATHEMATICS 2022. [DOI: 10.3390/math10152552] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Artificial intelligence (AI) is an evolving set of technologies used for solving a wide range of applied issues. The core of AI is machine learning (ML)—a complex of algorithms and methods that address the problems of classification, clustering, and forecasting. The practical application of AI&ML holds promising prospects. Therefore, the researches in this area are intensive. However, the industrial applications of AI and its more intensive use in society are not widespread at the present time. The challenges of widespread AI applications need to be considered from both the AI (internal problems) and the societal (external problems) perspective. This consideration will identify the priority steps for more intensive practical application of AI technologies, their introduction, and involvement in industry and society. The article presents the identification and discussion of the challenges of the employment of AI technologies in the economy and society of resource-based countries. The systematization of AI&ML technologies is implemented based on publications in these areas. This systematization allows for the specification of the organizational, personnel, social and technological limitations. This paper outlines the directions of studies in AI and ML, which will allow us to overcome some of the limitations and achieve expansion of the scope of AI&ML applications.
Collapse
|