251
|
Abstract
AbstractThis paper explores current developments in evolutionary and bio-inspired approaches to autonomous robotics, concentrating on research from our group at the University of Sussex. These developments are discussed in the context of advances in the wider fields of adaptive and evolutionary approaches to AI and robotics, focusing on the exploitation of embodied dynamics to create behaviour. Four case studies highlight various aspects of such exploitation. The first exploits the dynamical properties of a physical electronic substrate, demonstrating for the first time how component-level analog electronic circuits can be evolved directly in hardware to act as robot controllers. The second develops novel, effective and highly parsimonious navigation methods inspired by the way insects exploit the embodied dynamics of innate behaviours. Combining biological experiments with robotic modeling, it is shown how rapid route learning can be achieved with the aid of navigation-specific visual information that is provided and exploited by the innate behaviours. The third study focuses on the exploitation of neuromechanical chaos in the generation of robust motor behaviours. It is demonstrated how chaotic dynamics can be exploited to power a goal-driven search for desired motor behaviours in embodied systems using a particular control architecture based around neural oscillators. The dynamics are shown to be chaotic at all levels in the system, from the neural to the embodied mechanical. The final study explores the exploitation of the dynamics of brain-body-environment interactions for efficient, agile flapping winged flight. It is shown how a multi-objective evolutionary algorithm can be used to evolved dynamical neural controllers for a simulated flapping wing robot with feathered wings. Results demonstrate robust, stable, agile flight is achieved in the face of random wind gusts by exploiting complex asymmetric dynamics partly enabled by continually changing wing and tail morphologies.
Collapse
|
252
|
Robot navigation as hierarchical active inference. Neural Netw 2021; 142:192-204. [PMID: 34022669 DOI: 10.1016/j.neunet.2021.05.010] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2021] [Revised: 03/30/2021] [Accepted: 05/06/2021] [Indexed: 12/14/2022]
Abstract
Localization and mapping has been a long standing area of research, both in neuroscience, to understand how mammals navigate their environment, as well as in robotics, to enable autonomous mobile robots. In this paper, we treat navigation as inferring actions that minimize (expected) variational free energy under a hierarchical generative model. We find that familiar concepts like perception, path integration, localization and mapping naturally emerge from this active inference formulation. Moreover, we show that this model is consistent with models of hippocampal functions, and can be implemented in silico on a real-world robot. Our experiments illustrate that a robot equipped with our hierarchical model is able to generate topologically consistent maps, and correct navigation behaviour is inferred when a goal location is provided to the system.
Collapse
|
253
|
Zaffar M, Garg S, Milford M, Kooij J, Flynn D, McDonald-Maier K, Ehsan S. VPR-Bench: An Open-Source Visual Place Recognition Evaluation Framework with Quantifiable Viewpoint and Appearance Change. Int J Comput Vis 2021. [DOI: 10.1007/s11263-021-01469-5] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
AbstractVisual place recognition (VPR) is the process of recognising a previously visited place using visual information, often under varying appearance conditions and viewpoint changes and with computational constraints. VPR is related to the concepts of localisation, loop closure, image retrieval and is a critical component of many autonomous navigation systems ranging from autonomous vehicles to drones and computer vision systems. While the concept of place recognition has been around for many years, VPR research has grown rapidly as a field over the past decade due to improving camera hardware and its potential for deep learning-based techniques, and has become a widely studied topic in both the computer vision and robotics communities. This growth however has led to fragmentation and a lack of standardisation in the field, especially concerning performance evaluation. Moreover, the notion of viewpoint and illumination invariance of VPR techniques has largely been assessed qualitatively and hence ambiguously in the past. In this paper, we address these gaps through a new comprehensive open-source framework for assessing the performance of VPR techniques, dubbed “VPR-Bench”. VPR-Bench (Open-sourced at: https://github.com/MubarizZaffar/VPR-Bench) introduces two much-needed capabilities for VPR researchers: firstly, it contains a benchmark of 12 fully-integrated datasets and 10 VPR techniques, and secondly, it integrates a comprehensive variation-quantified dataset for quantifying viewpoint and illumination invariance. We apply and analyse popular evaluation metrics for VPR from both the computer vision and robotics communities, and discuss how these different metrics complement and/or replace each other, depending upon the underlying applications and system requirements. Our analysis reveals that no universal SOTA VPR technique exists, since: (a) state-of-the-art (SOTA) performance is achieved by 8 out of the 10 techniques on at least one dataset, (b) SOTA technique in one community does not necessarily yield SOTA performance in the other given the differences in datasets and metrics. Furthermore, we identify key open challenges since: (c) all 10 techniques suffer greatly in perceptually-aliased and less-structured environments, (d) all techniques suffer from viewpoint variance where lateral change has less effect than 3D change, and (e) directional illumination change has more adverse effects on matching confidence than uniform illumination change. We also present detailed meta-analyses regarding the roles of varying ground-truths, platforms, application requirements and technique parameters. Finally, VPR-Bench provides a unified implementation to deploy these VPR techniques, metrics and datasets, and is extensible through templates.
Collapse
|
254
|
Miao Z, Liu YH, Wang Y, Chen H, Zhong H, Fierro R. Consensus With Persistently Exciting Couplings and Its Application to Vision-Based Estimation. IEEE TRANSACTIONS ON CYBERNETICS 2021; 51:2801-2812. [PMID: 31180884 DOI: 10.1109/tcyb.2019.2918796] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
The problem of consensus in networked agent systems is revisited and applied to vision-based localization. A class of new consensus dynamics is introduced first, and sufficient conditions including the persistence of excitation on the coupling matrix for reaching consensus are derived. As an application of the proposed consensus dynamics, an adaptive localization algorithm then is proposed for autonomous robots equipped with primarily visual sensors in GPS-denied environments. In the context of consensus over an undirected tree topology, the convergence of the proposed localization algorithm is proved. Finally, both numerical simulations and physical experiments are presented to show the effectiveness of the proposed localization algorithm. Our algorithm is simpler to implement and computationally cheaper compared to other localization methods. Moreover, it is immune to error accumulation and long-term stable, and the asymptotical convergence of the estimation errors can be theoretically guaranteed.
Collapse
|
255
|
|
256
|
Hu X, Lang J. DOE-SLAM: Dynamic Object Enhanced Visual SLAM. SENSORS 2021; 21:s21093091. [PMID: 33946698 PMCID: PMC8124583 DOI: 10.3390/s21093091] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/30/2021] [Revised: 04/24/2021] [Accepted: 04/26/2021] [Indexed: 11/17/2022]
Abstract
In this paper, we formulate a novel strategy to adapt monocular-vision-based simultaneous localization and mapping (vSLAM) to dynamic environments. When enough background features can be captured, our system not only tracks the camera trajectory based on static background features but also estimates the foreground object motion from object features. In cases when a moving object obstructs too many background features for successful camera tracking from the background, our system can exploit the features from the object and the prediction of the object motion to estimate the camera pose. We use various synthetic and real-world test scenarios and the well-known TUM sequences to evaluate the capabilities of our system. The experiments show that we achieve higher pose estimation accuracy and robustness over state-of-the-art monocular vSLAM systems.
Collapse
|
257
|
Mapping in unstructured natural environment: a sensor fusion framework for wearable sensor suites. SN APPLIED SCIENCES 2021. [DOI: 10.1007/s42452-021-04555-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
AbstractWe present a generalized mapping framework that can withstand the challenges incurred by working in unstructured outdoor environments, such as a snowy forest. The proposed method takes advantage of a sensor fusion scheme, where sensors such as cameras and lidars are used in order to reconstruct the surrounding natural environment. Although mapping techniques such as SLAM and ICP cannot themselves properly handle the complexity of natural scenes, they do have the potential to contribute to the global solution in a proposed sensor fusion scheme, based on a factor graph architecture. In this paper, we propose an innovative map registration scheme for visual maps, and show how it can improve the reconstruction quality after data fusion. We also analyze the behavior and sensitivity of factor graphs to uncertainties, by comparing the residual error with different parameter combinations such as variances, using an exhaustive grid search with ground truth comparison. Finally, we suggest an ICP-inferred loop closure, capable of compensating position and attitude drift. The experiments are carried out by recording in a snowy forest using a wearable sensor suite. In the experiments, ground truth was acquired using a millimeter-accurate total station. The proposed framework is shown to be robust and likewise capable of providing estimates that are otherwise unattainable using classic techniques, such as visual SLAM and ICP for lasers. Finally, a visible improvement in the map reconstruction quality is shown, and the proposed framework achieves a translation error of 0.36 m.
Collapse
|
258
|
DARE-SLAM: Degeneracy-Aware and Resilient Loop Closing in Perceptually-Degraded Environments. J INTELL ROBOT SYST 2021. [DOI: 10.1007/s10846-021-01362-w] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
|
259
|
Morales J, Vázquez-Martín R, Mandow A, Morilla-Cabello D, García-Cerezo A. The UMA-SAR Dataset: Multimodal data collection from a ground vehicle during outdoor disaster response training exercises. Int J Rob Res 2021. [DOI: 10.1177/02783649211004959] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
This article presents a collection of multimodal raw data captured from a manned all-terrain vehicle in the course of two realistic outdoor search and rescue (SAR) exercises for actual emergency responders conducted in Málaga (Spain) in 2018 and 2019: the UMA-SAR dataset. The sensor suite, applicable to unmanned ground vehicles (UGVs), consisted of overlapping visible light (RGB) and thermal infrared (TIR) forward-looking monocular cameras, a Velodyne HDL-32 three-dimensional (3D) lidar, as well as an inertial measurement unit (IMU) and two global positioning system (GPS) receivers as ground truth. Our mission was to collect a wide range of data from the SAR domain, including persons, vehicles, debris, and SAR activity on unstructured terrain. In particular, four data sequences were collected following closed-loop routes during the exercises, with a total path length of 5.2 km and a total time of 77 min. In addition, we provide three more sequences of the empty site for comparison purposes (an extra 4.9 km and 46 min). Furthermore, the data is offered both in human-readable format and as rosbag files, and two specific software tools are provided for extracting and adapting this dataset to the users’ preference. The review of previously published disaster robotics repositories indicates that this dataset can contribute to fill a gap regarding visual and thermal datasets and can serve as a research tool for cross-cutting areas such as multispectral image fusion, machine learning for scene understanding, person and object detection, and localization and mapping in unstructured environments. The full dataset is publicly available at: www.uma.es/robotics-and-mechatronics/sar-datasets .
Collapse
Affiliation(s)
- Jesús Morales
- Universidad de Málaga, Andalucía Tech, Robotics and Mechatronics Group, Málaga, Spain
| | | | - Anthony Mandow
- Universidad de Málaga, Andalucía Tech, Robotics and Mechatronics Group, Málaga, Spain
| | - David Morilla-Cabello
- Universidad de Málaga, Andalucía Tech, Robotics and Mechatronics Group, Málaga, Spain
| | - Alfonso García-Cerezo
- Universidad de Málaga, Andalucía Tech, Robotics and Mechatronics Group, Málaga, Spain
| |
Collapse
|
260
|
Perico DH, Santos PE, Bianchi RAC. Guided navigation from multiple viewpoints using qualitative spatial reasoning. SPATIAL COGNITION AND COMPUTATION 2021. [DOI: 10.1080/13875868.2020.1857386] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Affiliation(s)
- D. H. Perico
- Department of Computer Science, Centro Universitáio FEI, São Bernardo Do Campo, Brazil
| | - P. E. Santos
- Department of Computer Science, Centro Universitáio FEI, São Bernardo Do Campo, Brazil
- College of Science and Engineering, Flinders University, Adelaide, Australia
| | - R. A. C. Bianchi
- Department of Computer Science, Centro Universitáio FEI, São Bernardo Do Campo, Brazil
| |
Collapse
|
261
|
Lluvia I, Lazkano E, Ansuategi A. Active Mapping and Robot Exploration: A Survey. SENSORS (BASEL, SWITZERLAND) 2021; 21:2445. [PMID: 33918107 PMCID: PMC8037480 DOI: 10.3390/s21072445] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/11/2021] [Revised: 03/21/2021] [Accepted: 03/28/2021] [Indexed: 11/16/2022]
Abstract
Simultaneous localization and mapping responds to the problem of building a map of the environment without any prior information and based on the data obtained from one or more sensors. In most situations, the robot is driven by a human operator, but some systems are capable of navigating autonomously while mapping, which is called native simultaneous localization and mapping. This strategy focuses on actively calculating the trajectories to explore the environment while building a map with a minimum error. In this paper, a comprehensive review of the research work developed in this field is provided, targeting the most relevant contributions in indoor mobile robotics.
Collapse
Affiliation(s)
- Iker Lluvia
- Autonomous and Intelligent Systems Unit, Fundación Tekniker, 20600 Eibar, Gipuzkoa, Spain;
| | - Elena Lazkano
- Robotics and Autonomous Systems Group (RSAIT), Computer Science and Artificial Intelligence Department, Faculty of Informatics, University of the Basque Country (UPV/EHU), 20018 Donostia, Gipuzkoa, Spain;
| | - Ander Ansuategi
- Autonomous and Intelligent Systems Unit, Fundación Tekniker, 20600 Eibar, Gipuzkoa, Spain;
| |
Collapse
|
262
|
Chen Y, Huang S, Zhao L, Dissanayake G. Cramér–Rao Bounds and Optimal Design Metrics for Pose-Graph SLAM. IEEE T ROBOT 2021. [DOI: 10.1109/tro.2020.3001718] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
263
|
Pham TH, Seto W, Daftry S, Ridge B, Hansen J, Thrush T, Van der Merwe M, Maggiolino G, Brinkman A, Mayo J, Cheng Y, Padgett C, Kulczycki E, Detry R. Rover Relocalization for Mars Sample Return by Virtual Template Synthesis and Matching. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3067281] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
264
|
Rodrigues RT, Tsiogkas N, Pascoal A, Aguiar AP. Online Range-Based SLAM Using B-Spline Surfaces. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3060672] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
265
|
Affiliation(s)
- Heng Yang
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Jingnan Shi
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Luca Carlone
- Laboratory for Information & Decision Systems (LIDS), Massachusetts Institute of Technology, Cambridge, MA, USA
| |
Collapse
|
266
|
Kalaitzakis M, Cain B, Carroll S, Ambrosi A, Whitehead C, Vitzilaios N. Fiducial Markers for Pose Estimation. J INTELL ROBOT SYST 2021. [DOI: 10.1007/s10846-020-01307-9] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
267
|
Azpúrua H, Rezende A, Potje G, Júnior GPDC, Fernandes R, Miranda V, Filho LWDR, Domingues J, Rocha F, de Sousa FLM, de Barros LGD, Nascimento ER, Macharet DG, Pessin G, Freitas GM. Towards Semi-autonomous Robotic Inspection and Mapping in Confined Spaces with the EspeleoRobô. J INTELL ROBOT SYST 2021. [DOI: 10.1007/s10846-021-01321-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
268
|
Tibebu H, Roche J, De Silva V, Kondoz A. LiDAR-Based Glass Detection for Improved Occupancy Grid Mapping. SENSORS (BASEL, SWITZERLAND) 2021; 21:2263. [PMID: 33804883 PMCID: PMC8038001 DOI: 10.3390/s21072263] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2021] [Revised: 03/19/2021] [Accepted: 03/21/2021] [Indexed: 12/01/2022]
Abstract
Creating an accurate awareness of the environment using laser scanners is a major challenge in robotics and auto industries. LiDAR (light detection and ranging) is a powerful laser scanner that provides a detailed map of the environment. However, efficient and accurate mapping of the environment is yet to be obtained, as most modern environments contain glass, which is invisible to LiDAR. In this paper, a method to effectively detect and localise glass using LiDAR sensors is proposed. This new approach is based on the variation of range measurements between neighbouring point clouds, using a two-step filter. The first filter examines the change in the standard deviation of neighbouring clouds. The second filter uses a change in distance and intensity between neighbouring pules to refine the results from the first filter and estimate the glass profile width before updating the cartesian coordinate and range measurement by the instrument. Test results demonstrate the detection and localisation of glass and the elimination of errors caused by glass in occupancy grid maps. This novel method detects frameless glass from a long range and does not depend on intensity peak with an accuracy of 96.2%.
Collapse
Affiliation(s)
- Haileleol Tibebu
- Institute of Digital Technologies, Loughborough University London, 3 Lesney Avenue, London E20 3BS, UK; (J.R.); (V.D.S.); (A.K.)
| | | | | | | |
Collapse
|
269
|
Abstract
This paper presents a novel bio-inspired predictive model of visual navigation inspired by mammalian navigation. This model takes inspiration from specific types of neurons observed in the brain, namely place cells, grid cells and head direction cells. In the proposed model, place cells are structures that store and connect local representations of the explored environment, grid and head direction cells make predictions based on these representations to define the position of the agent in a place cell’s reference frame. This specific use of navigation cells has three advantages: First, the environment representations are stored by place cells and require only a few spatialized descriptors or elements, making this model suitable for the integration of large-scale environments (indoor and outdoor). Second, the grid cell modules act as an efficient visual and absolute odometry system. Finally, the model provides sequential spatial tracking that can integrate and track an agent in redundant environments or environments with very few or no distinctive cues, while being very robust to environmental changes. This paper focuses on the architecture formalization and the main elements and properties of this model. The model has been successfully validated on basic functions: mapping, guidance, homing, and finding shortcuts. The precision of the estimated position of the agent and the robustness to environmental changes during navigation were shown to be satisfactory. The proposed predictive model is intended to be used on autonomous platforms, but also to assist visually impaired people in their mobility.
Collapse
|
270
|
Qu Y, Yang M, Zhang J, Xie W, Qiang B, Chen J. An Outline of Multi-Sensor Fusion Methods for Mobile Agents Indoor Navigation. SENSORS 2021; 21:s21051605. [PMID: 33668886 PMCID: PMC7956205 DOI: 10.3390/s21051605] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/16/2021] [Revised: 02/10/2021] [Accepted: 02/15/2021] [Indexed: 11/30/2022]
Abstract
Indoor autonomous navigation refers to the perception and exploration abilities of mobile agents in unknown indoor environments with the help of various sensors. It is the basic and one of the most important functions of mobile agents. In spite of the high performance of the single-sensor navigation method, multi-sensor fusion methods still potentially improve the perception and navigation abilities of mobile agents. This work summarizes the multi-sensor fusion methods for mobile agents’ navigation by: (1) analyzing and comparing the advantages and disadvantages of a single sensor in the task of navigation; (2) introducing the mainstream technologies of multi-sensor fusion methods, including various combinations of sensors and several widely recognized multi-modal sensor datasets. Finally, we discuss the possible technique trends of multi-sensor fusion methods, especially its technique challenges in practical navigation environments.
Collapse
Affiliation(s)
- Yuanhao Qu
- Research Center for Brain-inspired Intelligence (BII), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China;
- School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China; (J.Z.); (W.X.); (B.Q.); (J.C.)
| | - Minghao Yang
- Research Center for Brain-inspired Intelligence (BII), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China;
- Correspondence:
| | - Jiaqing Zhang
- School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China; (J.Z.); (W.X.); (B.Q.); (J.C.)
| | - Wu Xie
- School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China; (J.Z.); (W.X.); (B.Q.); (J.C.)
| | - Baohua Qiang
- School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China; (J.Z.); (W.X.); (B.Q.); (J.C.)
| | - Jinlong Chen
- School of Computer and Information Security, Guilin University of Electronic Technology, Guilin 541004, China; (J.Z.); (W.X.); (B.Q.); (J.C.)
| |
Collapse
|
271
|
Abstract
Sensing and mapping its surroundings is an essential requirement for a mobile robot. Geometric maps endow robots with the capacity of basic tasks, e.g., navigation. To co-exist with human beings in indoor scenes, the need to attach semantic information to a geometric map, which is called a semantic map, has been realized in the last two decades. A semantic map can help robots to behave in human rules, plan and perform advanced tasks, and communicate with humans on the conceptual level. This survey reviews methods about semantic mapping in indoor scenes. To begin with, we answered the question, what is a semantic map for mobile robots, by its definitions. After that, we reviewed works about each of the three modules of semantic mapping, i.e., spatial mapping, acquisition of semantic information, and map representation, respectively. Finally, though great progress has been made, there is a long way to implement semantic maps in advanced tasks for robots, thus challenges and potential future directions are discussed before a conclusion at last.
Collapse
|
272
|
Robust Visual-Inertial Navigation System for Low Precision Sensors under Indoor and Outdoor Environments. REMOTE SENSING 2021. [DOI: 10.3390/rs13040772] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Simultaneous Localization and Mapping (SLAM) has always been the focus of the robot navigation for many decades and becomes a research hotspot in recent years. Because a SLAM system based on vision sensor is vulnerable to environment illumination and texture, the problem of initial scale ambiguity still exists in a monocular SLAM system. The fusion of a monocular camera and an inertial measurement unit (IMU) can effectively solve the scale blur problem, improve the robustness of the system, and achieve higher positioning accuracy. Based on a monocular visual-inertial navigation system (VINS-mono), a state-of-the-art fusion performance of monocular vision and IMU, this paper designs a new initialization scheme that can calculate the acceleration bias as a variable during the initialization process so that it can be applied to low-cost IMU sensors. Besides, in order to obtain better initialization accuracy, visual matching positioning method based on feature point is used to assist the initialization process. After the initialization process, it switches to optical flow tracking visual positioning mode to reduce the calculation complexity. By using the proposed method, the advantages of feature point method and optical flow method can be fused. This paper, the first one to use both the feature point method and optical flow method, has better performance in the comprehensive performance of positioning accuracy and robustness under the low-cost sensors. Through experiments conducted with the EuRoc dataset and campus environment, the results show that the initial values obtained through the initialization process can be efficiently used for launching nonlinear visual-inertial state estimator and positioning accuracy of the improved VINS-mono has been improved by about 10% than VINS-mono.
Collapse
|
273
|
Earthen Jewish Architecture of Southern Morocco: Documentation of Unfired Brick Synagogues and Mellahs in the Drâa-Tafilalet Region. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11041712] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
This article seeks to highlight the vanished and not-so-well-known material culture of historical southern Moroccan Jewry. Jewish settlements could be found practically in the whole of North Africa before the Second World War; however, afterwards, it almost completely disappeared due to the political changes in the region and the establishment of the state of Israel. In southern Morocco, the last Jewish communities were present until the 1950s. Thanks to the interest of the Moroccan authorities, an effort has been made to restore some monuments and keep them as part of the cultural heritage that has attracted foreign tourists for the last few years. As part of the expeditionary research of Charles University and the Czech Technical University in Prague, several documentation projects were carried out in 2020, some of the results of which are described in this paper. Modern automatic methods of geomatics, such as easy to use laser scanning, mobile laser scanning in PLS modification (personal laser scanning), and close-range photogrammetry were used. The results of documentation were processed in the form of 3D models and basic plans, which are used mainly for analyzing residential zones of the Jewish population, the so-called mellahs. In this article, two case projects are described. In both cases, all the mentioned documentation methods were used. The technologies used were analyzed in terms of data collection speed, price, transport, and possible difficulties in use. The PLS technology is relatively new and still under development, such as miniaturising of other measuring instruments. Accuracy testing and usability of above-mentioned technology in cultural heritage documentation real practice is the benefit of this research. Finally, a second aim was to provide information of abandoned cultural places and constructions, which are on the edge of interest and endangered by destruction. It clearly shows that PLS technology is very fast and suitable for these types of objects.
Collapse
|
274
|
Arshad S, Kim GW. Role of Deep Learning in Loop Closure Detection for Visual and Lidar SLAM: A Survey. SENSORS 2021; 21:s21041243. [PMID: 33578695 PMCID: PMC7916334 DOI: 10.3390/s21041243] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/03/2020] [Revised: 01/27/2021] [Accepted: 02/04/2021] [Indexed: 11/16/2022]
Abstract
Loop closure detection is of vital importance in the process of simultaneous localization and mapping (SLAM), as it helps to reduce the cumulative error of the robot's estimated pose and generate a consistent global map. Many variations of this problem have been considered in the past and the existing methods differ in the acquisition approach of query and reference views, the choice of scene representation, and associated matching strategy. Contributions of this survey are many-fold. It provides a thorough study of existing literature on loop closure detection algorithms for visual and Lidar SLAM and discusses their insight along with their limitations. It presents a taxonomy of state-of-the-art deep learning-based loop detection algorithms with detailed comparison metrics. Also, the major challenges of conventional approaches are identified. Based on those challenges, deep learning-based methods were reviewed where the identified challenges are tackled focusing on the methods providing long-term autonomy in various conditions such as changing weather, light, seasons, viewpoint, and occlusion due to the presence of mobile objects. Furthermore, open challenges and future directions were also discussed.
Collapse
|
275
|
Li G, Zeng Y, Huang H, Song S, Liu B, Liao X. A Multi-Feature Fusion Slam System Attaching Semantic In-Variant to Points and Lines. SENSORS 2021; 21:s21041196. [PMID: 33567708 PMCID: PMC7916065 DOI: 10.3390/s21041196] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/05/2021] [Revised: 02/03/2021] [Accepted: 02/04/2021] [Indexed: 11/23/2022]
Abstract
The traditional simultaneous localization and mapping (SLAM) system uses static points of the environment as features for real-time localization and mapping. When there are few available point features, the system is difficult to implement. A feasible solution is to introduce line features. In complex scenarios containing rich line segments, the description of line segments is not strongly differentiated, which can lead to incorrect association of line segment data, thus introducing errors into the system and aggravating the cumulative error of the system. To address this problem, a point-line stereo visual SLAM system incorporating semantic invariants is proposed in this paper. This system improves the accuracy of line feature matching by fusing line features with image semantic invariant information. When defining the error function, the semantic invariant is fused with the reprojection error function, and the semantic constraint is applied to reduce the cumulative error of the poses in the long-term tracking process. Experiments on the Office sequence of the TartanAir dataset and the KITTI dataset show that this system improves the matching accuracy of line features and suppresses the cumulative error of the SLAM system to some extent, and the mean relative pose error (RPE) is 1.38 and 0.0593 m, respectively.
Collapse
Affiliation(s)
- Gang Li
- College of Electrical Engineering, Guangxi University, Nanning 530000, China; (G.L.); (Y.Z.); (S.S.); (B.L.); (X.L.)
| | - Yawen Zeng
- College of Electrical Engineering, Guangxi University, Nanning 530000, China; (G.L.); (Y.Z.); (S.S.); (B.L.); (X.L.)
| | - Huilan Huang
- College of Mechanical Engineering, Guangxi University, Nanning 530000, China
- Correspondence:
| | - Shaojian Song
- College of Electrical Engineering, Guangxi University, Nanning 530000, China; (G.L.); (Y.Z.); (S.S.); (B.L.); (X.L.)
| | - Bin Liu
- College of Electrical Engineering, Guangxi University, Nanning 530000, China; (G.L.); (Y.Z.); (S.S.); (B.L.); (X.L.)
- College of Automation, Central South University, Changsha 410083, China
| | - Xiang Liao
- College of Electrical Engineering, Guangxi University, Nanning 530000, China; (G.L.); (Y.Z.); (S.S.); (B.L.); (X.L.)
| |
Collapse
|
276
|
Miranda A, Hook JV, Schaal C. Lamb wave-based mapping of plate structures via frontier exploration. ULTRASONICS 2021; 110:106282. [PMID: 33142227 DOI: 10.1016/j.ultras.2020.106282] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/17/2020] [Revised: 09/22/2020] [Accepted: 10/09/2020] [Indexed: 06/11/2023]
Abstract
Substantial improvements in material processing and manufacturing techniques in recent years necessitate the introduction of effective and efficient nondestructive testing (NDT) methods that can seamlessly integrate into day-to-day aircraft and aerospace operations. Lamb wave-based methods have been identified as one of the most promising candidates for the inspection of large-scale structures. At the same time, there is presently a high level of research in the field of autonomous mobile robotics, especially in simultaneous localization and mapping (SLAM). Thus, this paper investigates a means to automate Lamb wave-based NDT by positioning sensors along a planar structure through mobile service robots. To this end, a generalized method for the mapping of plate structures using scattered Lamb waves by means of frontier exploration is presented such that an autonomous SLAM-capable NDT system can become realizable. The performance of this novel Lamb wave-based frontier exploration is first evaluated in simulation. It is shown that it generally outperforms a random frontier exploration and may even perform near-optimal in the case of an isotropic, square panel. These findings are then validated in laboratory experiments, confirming the general feasibility of utilizing Lamb waves for SLAM. Furthermore, the versatility of the developed methodology is successfully demonstrated on a more complexly shaped stiffened panel.
Collapse
Affiliation(s)
- Alvin Miranda
- Department of Mechanical Engineering, California State University, Northridge, CA, USA
| | - Joshua Vander Hook
- Maritime and Multi-Agent Autonomy Group, Jet Propulsion Laboratory, Pasadena, CA, USA
| | - Christoph Schaal
- Department of Mechanical Engineering, California State University, Northridge, CA, USA; Mechanical and Aerospace Engineering Department, University of California, Los Angeles, CA, USA.
| |
Collapse
|
277
|
Augmented Reality Based Surgical Navigation of Complex Pelvic Osteotomies—A Feasibility Study on Cadavers. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11031228] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
Abstract
Augmented reality (AR)-based surgical navigation may offer new possibilities for safe and accurate surgical execution of complex osteotomies. In this study we investigated the feasibility of navigating the periacetabular osteotomy of Ganz (PAO), known as one of the most complex orthopedic interventions, on two cadaveric pelves under realistic operating room conditions. Preoperative planning was conducted on computed tomography (CT)-reconstructed 3D models using an in-house developed software, which allowed creating cutting plane objects for planning of the osteotomies and reorientation of the acetabular fragment. An AR application was developed comprising point-based registration, motion compensation and guidance for osteotomies as well as fragment reorientation. Navigation accuracy was evaluated on CT-reconstructed 3D models, resulting in an error of 10.8 mm for osteotomy starting points and 5.4° for osteotomy directions. The reorientation errors were 6.7°, 7.0° and 0.9° for the x-, y- and z-axis, respectively. Average postoperative error of LCE angle was 4.5°. Our study demonstrated that the AR-based execution of complex osteotomies is feasible. Fragment realignment navigation needs further improvement, although it is more accurate than the state of the art in PAO surgery.
Collapse
|
278
|
Towards Semantic SLAM: 3D Position and Velocity Estimation by Fusing Image Semantic Information with Camera Motion Parameters for Traffic Scene Analysis. REMOTE SENSING 2021. [DOI: 10.3390/rs13030388] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
In this paper, an EKF (Extended Kalman Filter)-based algorithm is proposed to estimate 3D position and velocity components of different cars in a scene by fusing the semantic information and car model, extracted from successive frames with camera motion parameters. First, a 2D virtual image of the scene is made using a prior knowledge of the 3D Computer Aided Design (CAD) models of the detected cars and their predicted positions. Then, a discrepancy, i.e., distance, between the actual image and the virtual image is calculated. The 3D position and the velocity components are recursively estimated by minimizing the discrepancy using EKF. The experiments on the KiTTi dataset show a good performance of the proposed algorithm with a position estimation error up to 3–5% at 30 m and velocity estimation error up to 1 m/s.
Collapse
|
279
|
Yang GZ, Bellingham J, Dupont PE, Fischer P, Floridi L, Full R, Jacobstein N, Kumar V, McNutt M, Merrifield R, Nelson BJ, Scassellati B, Taddeo M, Taylor R, Veloso M, Wang ZL, Wood R. The grand challenges of Science Robotics. Sci Robot 2021; 3:3/14/eaar7650. [PMID: 33141701 DOI: 10.1126/scirobotics.aar7650] [Citation(s) in RCA: 359] [Impact Index Per Article: 119.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2017] [Accepted: 01/12/2018] [Indexed: 12/17/2022]
Abstract
One of the ambitions of Science Robotics is to deeply root robotics research in science while developing novel robotic platforms that will enable new scientific discoveries. Of our 10 grand challenges, the first 7 represent underpinning technologies that have a wider impact on all application areas of robotics. For the next two challenges, we have included social robotics and medical robotics as application-specific areas of development to highlight the substantial societal and health impacts that they will bring. Finally, the last challenge is related to responsible innovation and how ethics and security should be carefully considered as we develop the technology further.
Collapse
Affiliation(s)
- Guang-Zhong Yang
- Hamlyn Centre for Robotic Surgery, Imperial College London, London, UK.
| | - Jim Bellingham
- Center for Marine Robotics, Woods Hole Oceanographic Institution, Woods Hole, MA 02543, USA
| | - Pierre E Dupont
- Department of Cardiovascular Surgery, Boston Children's Hospital, Harvard Medical School, Boston, MA 02115, USA
| | - Peer Fischer
- Institute of Physical Chemistry, University of Stuttgart, Stuttgart, Germany.,Micro, Nano, and Molecular Systems Laboratory, Max Planck Institute for Intelligent Systems, Stuttgart, Germany
| | - Luciano Floridi
- Centre for Practical Ethics, Faculty of Philosophy, University of Oxford, Oxford, UK.,Digital Ethics Lab, Oxford Internet Institute, University of Oxford, Oxford, UK.,Department of Computer Science, University of Oxford, Oxford, UK.,Data Ethics Group, Alan Turing Institute, London, UK.,Department of Economics, American University, Washington, DC 20016, USA
| | - Robert Full
- Department of Integrative Biology, University of California, Berkeley, Berkeley, CA 94720, USA
| | - Neil Jacobstein
- Singularity University, NASA Research Park, Moffett Field, CA 94035, USA.,MediaX, Stanford University, Stanford, CA 94305, USA
| | - Vijay Kumar
- Department of Mechanical Engineering and Applied Mechanics, University of Pennsylvania, Philadelphia, PA 19104, USA
| | - Marcia McNutt
- National Academy of Sciences, Washington, DC 20418, USA
| | - Robert Merrifield
- Hamlyn Centre for Robotic Surgery, Imperial College London, London, UK
| | - Bradley J Nelson
- Institute of Robotics and Intelligent Systems, Department of Mechanical and Process Engineering, ETH Zürich, Zurich, Switzerland
| | - Brian Scassellati
- Department of Computer Science, Yale University, New Haven, CT 06520, USA.,Department Mechanical Engineering and Materials Science, Yale University, New Haven, CT 06520, USA
| | - Mariarosaria Taddeo
- Digital Ethics Lab, Oxford Internet Institute, University of Oxford, Oxford, UK.,Department of Computer Science, University of Oxford, Oxford, UK.,Data Ethics Group, Alan Turing Institute, London, UK
| | - Russell Taylor
- Department of Computer Science, Johns Hopkins University, Baltimore, MD 21218, USA
| | - Manuela Veloso
- Machine Learning Department, School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213, USA
| | - Zhong Lin Wang
- School of Materials Science and Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
| | - Robert Wood
- John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, MA 02138, USA.,Wyss Institute for Biologically Inspired Engineering, Harvard University, Cambridge, MA 02138, USA
| |
Collapse
|
280
|
Detection of False Synchronization of Stereo Image Transmission Using a Convolutional Neural Network. Symmetry (Basel) 2021. [DOI: 10.3390/sym13010078] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The subject of the work described in this article is the detection of false synchronization in the transmission of digital stereo images. Until now, the synchronization problem was solved by using start triggers in the recording. Our proposal checks the discrepancy between the received pairs of images, which allows you to detect delays in transferring images between the left camera and the right camera. For this purpose, a deep network is used to classify the analyzed pairs of images into five classes: MuchFaster, Faster, Regular, Slower, and MuchSlower. As can be seen as a result of the conducted work, satisfactory research results were obtained as the correct classification. A high percentage of average probability in individual classes also indicates a high degree of certainty as to the correctness of the results. An author’s base of colorful stereo images in the number of 3070 pairs is used for the research.
Collapse
|
281
|
Chen L, Jin S, Xia Z. Towards a Robust Visual Place Recognition in Large-Scale vSLAM Scenarios Based on a Deep Distance Learning. SENSORS 2021; 21:s21010310. [PMID: 33466401 PMCID: PMC7796086 DOI: 10.3390/s21010310] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/16/2020] [Revised: 12/31/2020] [Accepted: 01/04/2021] [Indexed: 11/16/2022]
Abstract
The application of deep learning is blooming in the field of visual place recognition, which plays a critical role in visual Simultaneous Localization and Mapping (vSLAM) applications. The use of convolutional neural networks (CNNs) achieve better performance than handcrafted feature descriptors. However, visual place recognition is still a challenging task due to two major problems, i.e., perceptual aliasing and perceptual variability. Therefore, designing a customized distance learning method to express the intrinsic distance constraints in the large-scale vSLAM scenarios is of great importance. Traditional deep distance learning methods usually use the triplet loss which requires the mining of anchor images. This may, however, result in very tedious inefficient training and anomalous distance relationships. In this paper, a novel deep distance learning framework for visual place recognition is proposed. Through in-depth analysis of the multiple constraints of the distance relationship in the visual place recognition problem, the multi-constraint loss function is proposed to optimize the distance constraint relationships in the Euclidean space. The new framework can support any kind of CNN such as AlexNet, VGGNet and other user-defined networks to extract more distinguishing features. We have compared the results with the traditional deep distance learning method, and the results show that the proposed method can improve the performance by 19–28%. Additionally, compared to some contemporary visual place recognition techniques, the proposed method can improve the performance by 40%/36% and 27%/24% in average on VGGNet/AlexNet using the New College and the TUM datasets, respectively. It’s verified the method is capable to handle appearance changes in complex environments.
Collapse
Affiliation(s)
- Liang Chen
- Correspondence: ; Tel.: +86-185-5040-8581
| | | | | |
Collapse
|
282
|
Particle swarm optimization for solving a scan-matching problem based on the normal distributions transform. EVOLUTIONARY INTELLIGENCE 2021. [DOI: 10.1007/s12065-020-00545-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
283
|
Omranpour H, Shiry S. A mobile robot mapping model inspired from the place cells functionality of hippocampus based on dimension reduction technique. J EXP THEOR ARTIF IN 2021. [DOI: 10.1080/0952813x.2020.1721569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Affiliation(s)
- Hesam Omranpour
- Electrical and Computer Engineering Department, Babol Noshirvani University of Technology, Babol, Iran
| | - Saeed Shiry
- Computer Engineering Department, Amirkabir University of Technology, Tehran, Iran
| |
Collapse
|
284
|
Wang K, Ma S, Chen J, Ren F, Lu J. Approaches Challenges and Applications for Deep Visual Odometry Toward to Complicated and Emerging Areas. IEEE Trans Cogn Dev Syst 2021. [DOI: 10.1109/tcds.2020.3038898] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
285
|
He J, Zhou Y, Huang L, Kong Y, Cheng H. Ground and Aerial Collaborative Mapping in Urban Environments. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2020.3032054] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
286
|
Jiao J, Ye H, Zhu Y, Liu M. Robust Odometry and Mapping for Multi-LiDAR Systems With Online Extrinsic Calibration. IEEE T ROBOT 2021. [DOI: 10.1109/tro.2021.3078287] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
287
|
Zempo K, Arai T, Aoki T, Okada Y. Sensing Framework for the Internet of Actors in the Value Co-Creation Process with a Beacon-Attachable Indoor Positioning System. SENSORS 2020; 21:s21010083. [PMID: 33375596 PMCID: PMC7795509 DOI: 10.3390/s21010083] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/17/2020] [Revised: 12/19/2020] [Accepted: 12/22/2020] [Indexed: 01/10/2023]
Abstract
To evaluate and improve the value of a service, it is important to measure not only the outcomes, but also the process of the service. Value co-creation (VCC) is not limited to outcomes, especially in interpersonal services based on interactions between actors. In this paper, a sensing framework for a VCC process in retail stores is proposed by improving an environment recognition based indoor positioning system with high positioning performance in a metal shelf environment. The conventional indoor positioning systems use radio waves; therefore, errors are caused by reflection, absorption, and interference from metal shelves. An improvement in positioning performance was achieved in the proposed method by using an IR (infrared) slit and IR light, which avoids such errors. The system was designed to recognize many and unspecified people based on the environment recognition method that the receivers had installed, in the service environment. In addition, sensor networking was also conducted by adding a function to transmit payload and identification simultaneously to the beacons that were attached to positioning objects. The effectiveness of the proposed method was verified by installing it not only in an experimental environment with ideal conditions, but posteriorly, the system was tested in real conditions, in a retail store. In our experimental setup, in a comparison with equal element numbers, positioning identification was possible within an error of 96.2 mm in a static environment in contrast to the radio wave based method where an average positioning error of approximately 648 mm was measured using the radio wave based method (Bluetooth low-energy fingerprinting technique). Moreover, when multiple beacons were used simultaneously in our system within the measurement range of one receiver, the appropriate setting of the pulse interval and jitter rate was implemented by simulation. Additionally, it was confirmed that, in a real scenario, it is possible to measure the changes in movement and positional relationships between people. This result shows the feasibility of measuring and evaluating the VCC process in retail stores, although it was difficult to measure the interaction between actors.
Collapse
Affiliation(s)
- Keiichi Zempo
- Faculty of Engineering, Information and Systems, University of Tsukuba, Tsukuba 305-8573, Ibaraki, Japan;
- Correspondence:
| | - Taiga Arai
- Graduate School of Systems and Information Engineering, University of Tsukuba, Tsukuba 305-8573, Ibaraki, Japan; (T.A.); (T.A.)
| | - Takuya Aoki
- Graduate School of Systems and Information Engineering, University of Tsukuba, Tsukuba 305-8573, Ibaraki, Japan; (T.A.); (T.A.)
| | - Yukihiko Okada
- Faculty of Engineering, Information and Systems, University of Tsukuba, Tsukuba 305-8573, Ibaraki, Japan;
| |
Collapse
|
288
|
Mascaro R, Wermelinger M, Hutter M, Chli M. Towards automating construction tasks: Large‐scale object mapping, segmentation, and manipulation. J FIELD ROBOT 2020. [DOI: 10.1002/rob.22007] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Ruben Mascaro
- Vision for Robotics Lab (V4RL) ETH Zurich Zurich Switzerland
| | | | - Marco Hutter
- Robotic Systems Lab (RSL) ETH Zurich Zurich Switzerland
| | - Margarita Chli
- Vision for Robotics Lab (V4RL) ETH Zurich Zurich Switzerland
| |
Collapse
|
289
|
Zhang X, Liu Q, Zheng B, Wang H, Wang Q. A visual simultaneous localization and mapping approach based on scene segmentation and incremental optimization. INT J ADV ROBOT SYST 2020. [DOI: 10.1177/1729881420977669] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
Existing visual simultaneous localization and mapping (V-SLAM) algorithms are usually sensitive to the situation with sparse landmarks in the environment and large view transformation of camera motion, and they are liable to generate large pose errors that lead to track failures due to the decrease of the matching rate of feature points. Aiming at the above problems, this article proposes an improved V-SLAM method based on scene segmentation and incremental optimization strategy. In the front end, this article proposes a scene segmentation algorithm considering camera motion direction and angle. By segmenting the trajectory and adding camera motion direction to the tracking thread, an effective prediction model of camera motion in the scene with sparse landmarks and large view transformation is realized. In the back end, this article proposes an incremental optimization method combining segmentation information and an optimization method for tracking prediction model. By incrementally adding the state parameters and reusing the computed results, high-precision results of the camera trajectory and feature points are obtained with satisfactory computing speed. The performance of our algorithm is evaluated by two well-known datasets: TUM RGB-D and NYUDv2 RGB-D. The experimental results demonstrate that our method improves the computational efficiency by 10.2% compared with state-of-the-art V-SLAMs on the desktop platform and by 22.4% on the embedded platform, respectively. Meanwhile, the robustness of our method is better than that of ORB-SLAM2 on the TUM RGB-D dataset.
Collapse
Affiliation(s)
- Xiaoguo Zhang
- School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Qihan Liu
- School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Bingqing Zheng
- School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Huiqing Wang
- School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Qing Wang
- School of Instrument Science and Engineering, Southeast University, Nanjing, China
| |
Collapse
|
290
|
Navigation and Mapping in Forest Environment Using Sparse Point Clouds. REMOTE SENSING 2020. [DOI: 10.3390/rs12244088] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Odometry during forest operations is demanding, involving limited field of vision (FOV), back-and-forth work cycle movements, and occasional close obstacles, which create problems for state-of-the-art systems. We propose a two-phase on-board process, where tree stem registration produces a sparse point cloud (PC) which is then used for simultaneous location and mapping (SLAM). A field test was carried out using a harvester with a laser scanner and a global navigation satellite system (GNSS) performing forest thinning over a 520 m strip route. Two SLAM methods are used: The proposed sparse SLAM (sSLAM) and a standard method, LeGO-LOAM (LLOAM). A generic SLAM post-processing method is presented, which improves the odometric accuracy with a small additional processing cost. The sSLAM method uses only tree stem centers, reducing the allocated memory to approximately 1% of the total PC size. Odometry and mapping comparisons between sSLAM and LLOAM are presented. Both methods show 85% agreement in registration within 15 m of the strip road and odometric accuracy of 0.5 m per 100 m. Accuracy is evaluated by comparing the harvester location derived through odometry to locations collected by a GNSS receiver mounted on the harvester.
Collapse
|
291
|
Fan T, Wang H, Rubenstein M, Murphey T. CPL-SLAM: Efficient and Certifiably Correct Planar Graph-Based SLAM Using the Complex Number Representation. IEEE T ROBOT 2020. [DOI: 10.1109/tro.2020.3006717] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
292
|
|
293
|
|
294
|
Abstract
In this paper, we formulate the active SLAM paradigm in terms of model-free Deep Reinforcement Learning, embedding the traditional utility functions based on the Theory of Optimal Experimental Design in rewards, and therefore relaxing the intensive computations of classical approaches. We validate such formulation in a complex simulation environment, using a state-of-the-art deep Q-learning architecture with laser measurements as network inputs. Trained agents become capable not only to learn a policy to navigate and explore in the absence of an environment model but also to transfer their knowledge to previously unseen maps, which is a key requirement in robotic exploration.
Collapse
|
295
|
Abstract
Research and development of autonomous mobile robotic solutions that can perform several active agricultural tasks (pruning, harvesting, mowing) have been growing. Robots are now used for a variety of tasks such as planting, harvesting, environmental monitoring, supply of water and nutrients, and others. To do so, robots need to be able to perform online localization and, if desired, mapping. The most used approach for localization in agricultural applications is based in standalone Global Navigation Satellite System-based systems. However, in many agricultural and forest environments, satellite signals are unavailable or inaccurate, which leads to the need of advanced solutions independent from these signals. Approaches like simultaneous localization and mapping and visual odometry are the most promising solutions to increase localization reliability and availability. This work leads to the main conclusion that, few methods can achieve simultaneously the desired goals of scalability, availability, and accuracy, due to the challenges imposed by these harsh environments. In the near future, novel contributions to this field are expected that will help one to achieve the desired goals, with the development of more advanced techniques, based on 3D localization, and semantic and topological mapping. In this context, this work proposes an analysis of the current state-of-the-art of localization and mapping approaches in agriculture and forest environments. Additionally, an overview about the available datasets to develop and test these approaches is performed. Finally, a critical analysis of this research field is done, with the characterization of the literature using a variety of metrics.
Collapse
|
296
|
Abstract
SUMMARYSelf-localization in highly dynamic environments is still a challenging problem for humanoid robots with limited computation resource. In this paper, we propose a dual-channel unscented particle filter (DC-UPF)-based localization method to address it. A key novelty of this approach is that it employs a dual-channel switch mechanism in measurement updating procedure of particle filter, solving for sparse vision feature in motion, and it leverages data from a camera, a walking odometer, and an inertial measurement unit. Extensive experiments with an NAO robot demonstrate that DC-UPF outperforms UPF and Monte–Carlo localization with regard to accuracy.
Collapse
|
297
|
Tawiah TAQ. A review of algorithms and techniques for image-based recognition and inference in mobile robotic systems. INT J ADV ROBOT SYST 2020. [DOI: 10.1177/1729881420972278] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
Autonomous vehicles include driverless, self-driving and robotic cars, and other platforms capable of sensing and interacting with its environment and navigating without human help. On the other hand, semiautonomous vehicles achieve partial realization of autonomy with human intervention, for example, in driver-assisted vehicles. Autonomous vehicles first interact with their surrounding using mounted sensors. Typically, visual sensors are used to acquire images, and computer vision techniques, signal processing, machine learning, and other techniques are applied to acquire, process, and extract information. The control subsystem interprets sensory information to identify appropriate navigation path to its destination and action plan to carry out tasks. Feedbacks are also elicited from the environment to improve upon its behavior. To increase sensing accuracy, autonomous vehicles are equipped with many sensors [light detection and ranging (LiDARs), infrared, sonar, inertial measurement units, etc.], as well as communication subsystem. Autonomous vehicles face several challenges such as unknown environments, blind spots (unseen views), non-line-of-sight scenarios, poor performance of sensors due to weather conditions, sensor errors, false alarms, limited energy, limited computational resources, algorithmic complexity, human–machine communications, size, and weight constraints. To tackle these problems, several algorithmic approaches have been implemented covering design of sensors, processing, control, and navigation. The review seeks to provide up-to-date information on the requirements, algorithms, and main challenges in the use of machine vision–based techniques for navigation and control in autonomous vehicles. An application using land-based vehicle as an Internet of Thing-enabled platform for pedestrian detection and tracking is also presented.
Collapse
|
298
|
Zuo G, Zheng T, Liu Y, Xu Z, Gong D, Yu J. Fine semantic mapping based on dense segmentation network. INTEL SERV ROBOT 2020. [DOI: 10.1007/s11370-020-00341-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
299
|
Yang S, Li B, Liu M, Lai YK, Kobbelt L, Hu SM. HeteroFusion: Dense Scene Reconstruction Integrating Multi-Sensors. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2020; 26:3217-3230. [PMID: 31150341 DOI: 10.1109/tvcg.2019.2919619] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
We present a novel approach to integrate data from multiple sensor types for dense 3D reconstruction of indoor scenes in realtime. Existing algorithms are mainly based on a single RGBD camera and thus require continuous scanning of areas with sufficient geometric features. Otherwise, tracking may fail due to unreliable frame registration. Inspired by the fact that the fusion of multiple sensors can combine their strengths towards a more robust and accurate self-localization, we incorporate multiple types of sensors which are prevalent in modern robot systems, including a 2D range sensor, an inertial measurement unit (IMU), and wheel encoders. We fuse their measurements to reinforce the tracking process and to eventually obtain better 3D reconstructions. Specifically, we develop a 2D truncated signed distance field (TSDF) volume representation for the integration and ray-casting of laser frames, leading to a unified cost function in the pose estimation stage. For validation of the estimated poses in the loop-closure optimization process, we train a classifier for the features extracted from heterogeneous sensors during the registration progress. To evaluate our method on challenging use case scenarios, we assembled a scanning platform prototype to acquire real-world scans. We further simulated synthetic scans based on high-fidelity synthetic scenes for quantitative evaluation. Extensive experimental evaluation on these two types of scans demonstrate that our system is capable of robustly acquiring dense 3D reconstructions and outperforms state-of-the-art RGBD and LiDAR systems.
Collapse
|
300
|
Sleaman WK, Yavuz S. Indoor mobile robot navigation using deep convolutional neural network. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2020. [DOI: 10.3233/jifs-189030] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Robot can help human in their everyday life and routine. These are not an indoor robot which was designed to perform desired task, but they can adapt to our environment by themselves and to learn from their own experiences. In this research we focus on high degree of autonomy, which is a must for social robots. For training purpose autonomous exploration and unknown environments is used along with proper algorithm so that robot can adapt to unknown environments. For testing purpose, simulation is carried with sensor fusion method, so that real world noise can be reduced and accuracy can be increased. This dissertation focuses on the intelligent robot control in autonomous navigation tasks and investigates the robot learning in following aspects. This method is based on human instinct of imitation. In this standard real time data set is provided to the robot for training purpose, it gets train from these data and generalize over all unseen potential situations and environments. Convolutional Neural Network is used to determine the probability and based on that robot can act. After acceptable number of demonstrations, robot can predict output with high accuracy and hence can acquire the independent navigation skills. State-of-the-art reinforcement learning techniques is used to train the robot via interaction with the robots. Convolutional Neural Network is also incorporated for fast generalization. Robot is train based on all past state-action pairs collected during interaction. This training model can predict output which helps robot for autonomous navigation.
Collapse
Affiliation(s)
- Walead Kaled Sleaman
- Faculty of computer and mathematic, Department of Computer, Tikrit University, Turkey
| | - Sırma Yavuz
- Faculty of Electrical and Electronics, Department of Computer Engineering, Yildiz Technical University, Turkey
| |
Collapse
|