401
|
Yang B, Wang S, Markham A, Trigoni N. Robust Attentional Aggregation of Deep Feature Sets for Multi-view 3D Reconstruction. Int J Comput Vis 2019. [DOI: 10.1007/s11263-019-01217-w] [Citation(s) in RCA: 36] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
402
|
SLAM in Dynamic Environments: A Deep Learning Approach for Moving Object Tracking Using ML-RANSAC Algorithm. SENSORS 2019; 19:s19173699. [PMID: 31454925 PMCID: PMC6749210 DOI: 10.3390/s19173699] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/25/2019] [Revised: 08/16/2019] [Accepted: 08/18/2019] [Indexed: 11/17/2022]
Abstract
The important problem of Simultaneous Localization and Mapping (SLAM) in dynamic environments is less studied than the counterpart problem in static settings. In this paper, we present a solution for the feature-based SLAM problem in dynamic environments. We propose an algorithm that integrates SLAM with multi-target tracking (SLAMMTT) using a robust feature-tracking algorithm for dynamic environments. A novel implementation of RANdomSAmple Consensus (RANSAC) method referred to as multilevel-RANSAC (ML-RANSAC) within the Extended Kalman Filter (EKF) framework is applied for multi-target tracking (MTT). We also apply machine learning to detect features from the input data and to distinguish moving from stationary objects. The data stream from LIDAR and vision sensors are fused in real-time to detect objects and depth information. A practical experiment is designed to verify the performance of the algorithm in a dynamic environment. The unique feature of this algorithm is its ability to maintain tracking of features even when the observations are intermittent whereby many reported algorithms fail in such situations. Experimental validation indicates that the algorithm is able to perform consistent estimates in a fast and robust manner suggesting its feasibility for real-time applications.
Collapse
|
403
|
Nowicki MR, Skrzypczyński P. Leveraging Visual Place Recognition to Improve Indoor Positioning with Limited Availability of WiFi Scans. SENSORS 2019; 19:s19173657. [PMID: 31443504 PMCID: PMC6749292 DOI: 10.3390/s19173657] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/17/2019] [Revised: 08/15/2019] [Accepted: 08/20/2019] [Indexed: 11/16/2022]
Abstract
WiFi-based fingerprinting is promising for practical indoor localization with smartphones because this technique provides absolute estimates of the current position, while the WiFi infrastructure is ubiquitous in the majority of indoor environments. However, the application of WiFi fingerprinting for positioning requires pre-surveyed signal maps and is getting more restricted in the recent generation of smartphones due to changes in security policies. Therefore, we sought new sources of information that can be fused into the existing indoor positioning framework, helping users to pinpoint their position, even with a relatively low-quality, sparse WiFi signal map. In this paper, we demonstrate that such information can be derived from the recognition of camera images. We present a way of transforming qualitative information of image similarity into quantitative constraints that are then fused into the graph-based optimization framework for positioning together with typical pedestrian dead reckoning (PDR) and WiFi fingerprinting constraints. Performance of the improved indoor positioning system is evaluated on different user trajectories logged inside an office building at our University campus. The results demonstrate that introducing additional sensing modality into the positioning system makes it possible to increase accuracy and simultaneously reduce the dependence on the quality of the pre-surveyed WiFi map and the WiFi measurements at run-time.
Collapse
Affiliation(s)
- Michał R Nowicki
- Institute of Control, Robotics and Information Engineering, Poznan University of Technology, 60-965 Poznan, Poland.
| | - Piotr Skrzypczyński
- Institute of Control, Robotics and Information Engineering, Poznan University of Technology, 60-965 Poznan, Poland
| |
Collapse
|
404
|
Delmerico J, Mintchev S, Giusti A, Gromov B, Melo K, Horvat T, Cadena C, Hutter M, Ijspeert A, Floreano D, Gambardella LM, Siegwart R, Scaramuzza D. The current state and future outlook of rescue robotics. J FIELD ROBOT 2019. [DOI: 10.1002/rob.21887] [Citation(s) in RCA: 102] [Impact Index Per Article: 20.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Jeffrey Delmerico
- Robotics and Perception Group, Department of Informatics and NeuroinformaticsUniversity of Zurich and ETH, Zurich Zürich Switzerland
| | - Stefano Mintchev
- Laboratory of Intelligent SystemsSwiss Federal Institute of Technology Lausanne Switzerland
| | - Alessandro Giusti
- Dalle Molle Institute for Artificial Intelligence (IDSIA), USI‐SUPSI Manno Switzerland
| | - Boris Gromov
- Dalle Molle Institute for Artificial Intelligence (IDSIA), USI‐SUPSI Manno Switzerland
| | - Kamilo Melo
- Biorobotics LaboratorySwiss Federal Institute of Technology Lausanne Switzerland
| | - Tomislav Horvat
- Biorobotics LaboratorySwiss Federal Institute of Technology Lausanne Switzerland
| | - Cesar Cadena
- Autonomous Systems LabSwiss Federal Institute of Technology Zürich Switzerland
| | - Marco Hutter
- Robotic Systems LabSwiss Federal Institute of Technology Zürich Switzerland
| | - Auke Ijspeert
- Biorobotics LaboratorySwiss Federal Institute of Technology Lausanne Switzerland
| | - Dario Floreano
- Laboratory of Intelligent SystemsSwiss Federal Institute of Technology Lausanne Switzerland
| | - Luca M. Gambardella
- Dalle Molle Institute for Artificial Intelligence (IDSIA), USI‐SUPSI Manno Switzerland
| | - Roland Siegwart
- Autonomous Systems LabSwiss Federal Institute of Technology Zürich Switzerland
| | - Davide Scaramuzza
- Robotics and Perception Group, Department of Informatics and NeuroinformaticsUniversity of Zurich and ETH, Zurich Zürich Switzerland
| |
Collapse
|
405
|
A Low Overhead Mapping Scheme for Exploration and Representation in the Unknown Area. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9153089] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The grid map, representing area information with the number of cells, is a widely used mapping scheme for mobile robots and simultaneous localization and mapping (SLAM) processes. However, the tremendous amount of cells in a grid map for a detailed map representation results in overheads for memory space and computing paths in mobile robots. Therefore, to overcome the overhead of the grid map, this study proposes a new low overhead mapping scheme which the authors call as the Rmap that represents an area with variable sizes of rectangles instead of the number of cells in the grid map. This mapping scheme also provides an exploration path for obtaining new information for the unknown area. This study evaluated the performance of the Rmap in real environments as well as in simulation environments. The experiment results show that the Rmap can reduce the overhead of a grid map. In one of our experimental environments, the Rmap represented an area with 85% less memory than the grid map. The Rmap also showed better coverage performance compared with other previous algorithms.
Collapse
|
406
|
Abstract
The high-definition map (HD-map) of road structures is crucial for the safe planning and control of autonomous vehicles. However, generating and updating such maps requires intensive manual work. Simultaneous localization and mapping (SLAM) is able to automatically build and update a map of the environment. Nevertheless, there is still a lack of SLAM method for generating vector-based road structure maps. In this paper, we propose a vector-based SLAM method for the road structure mapping using vehicle-mounted multibeam LiDAR. We propose using polylines as the primary mapping element instead of grid maps or point clouds because the vector-based representation is lightweight and precise. We explored the following: (1) the extraction and vectorization of road structures based on multiframe probabilistic fusion; (2) the efficient vector-based matching between frames of road structures; (3) the loop closure and optimization based on the pose-graph; and (4) the global reconstruction of the vector map. One specific road structure, the road boundary, is taken as an example. We applied the proposed mapping method to three road scenes, ranging from hundreds of meters to over ten kilometers and the results are automatically generated vector-based road boundary maps. The average absolute pose error of the trajectory in the mapping is 1.83 m without the aid of high-precision GPS.
Collapse
|
407
|
Dubé R, Cramariuc A, Dugas D, Sommer H, Dymczyk M, Nieto J, Siegwart R, Cadena C. SegMap: Segment-based mapping and localization using data-driven descriptors. Int J Rob Res 2019. [DOI: 10.1177/0278364919863090] [Citation(s) in RCA: 56] [Impact Index Per Article: 11.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Precisely estimating a robot’s pose in a prior, global map is a fundamental capability for mobile robotics, e.g., autonomous driving or exploration in disaster zones. This task, however, remains challenging in unstructured, dynamic environments, where local features are not discriminative enough and global scene descriptors only provide coarse information. We therefore present SegMap: a map representation solution for localization and mapping based on the extraction of segments in 3D point clouds. Working at the level of segments offers increased invariance to view-point and local structural changes, and facilitates real-time processing of large-scale 3D data. SegMap exploits a single compact data-driven descriptor for performing multiple tasks: global localization, 3D dense map reconstruction, and semantic information extraction. The performance of SegMap is evaluated in multiple urban driving and search and rescue experiments. We show that the learned SegMap descriptor has superior segment retrieval capabilities, compared with state-of-the-art handcrafted descriptors. As a consequence, we achieve a higher localization accuracy and a 6% increase in recall over state-of-the-art handcrafted descriptors. These segment-based localizations allow us to reduce the open-loop odometry drift by up to 50%. SegMap is open-source available along with easy to run demonstrations.
Collapse
Affiliation(s)
- Renaud Dubé
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
- Sevensense Robotics AG, Zurich, Switzerland
| | | | - Daniel Dugas
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
| | - Hannes Sommer
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
- Sevensense Robotics AG, Zurich, Switzerland
| | - Marcin Dymczyk
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
- Sevensense Robotics AG, Zurich, Switzerland
| | - Juan Nieto
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
| | | | - Cesar Cadena
- Autonomous Systems Lab (ASL), ETH Zurich, Switzerland
| |
Collapse
|
408
|
Aloise I, Corte BD, Nardi F, Grisetti G. Systematic Handling of Heterogeneous Geometric Primitives in Graph-SLAM Optimization. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2918054] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
409
|
Zheng F, Tang H, Liu YH. Odometry-Vision-Based Ground Vehicle Motion Estimation With SE(2)-Constrained SE(3) Poses. IEEE TRANSACTIONS ON CYBERNETICS 2019; 49:2652-2663. [PMID: 29993766 DOI: 10.1109/tcyb.2018.2831900] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
This paper focuses on the motion estimation problem of ground vehicles using odometry and monocular visual sensors. While the keyframe-based batch optimization methods become the mainstream approach in mobile vehicle localization and mapping, the keyframe poses are usually represented by SE(3) in vision-based methods or SE(2) in methods based on range scanners. For a ground vehicle, this paper proposes a new SE(2)-constrained SE(3) parameterization of its poses, which can be easily achieved in the batch optimization framework using specially formulated edges. Utilizing such a parameterization of poses, a complete odometry-vision-based motion estimation system is developed. The system is designed in a commonly used structure of graph optimization, providing high modularity and flexibility for further implementation or adaptation. Its superior performance in terms of accuracy on a ground vehicle platform is validated by real-world experiments in industrial indoor environments.
Collapse
|
410
|
Lahemer ES, Rad A. An Adaptive Augmented Vision-Based Ellipsoidal SLAM for Indoor Environments. SENSORS 2019; 19:s19122795. [PMID: 31234441 PMCID: PMC6630515 DOI: 10.3390/s19122795] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/03/2019] [Revised: 06/17/2019] [Accepted: 06/19/2019] [Indexed: 11/22/2022]
Abstract
In this paper, the problem of Simultaneous Localization And Mapping (SLAM) is addressed via a novel augmented landmark vision-based ellipsoidal SLAM. The algorithm is implemented on a NAO humanoid robot and is tested in an indoor environment. The main feature of the system is the implementation of SLAM with a monocular vision system. Distinguished landmarks referred to as NAOmarks are employed to localize the robot via its monocular vision system. We henceforth introduce the notion of robotic augmented reality (RAR) and present a monocular Extended Kalman Filter (EKF)/ellipsoidal SLAM in order to improve the performance and alleviate the computational effort, to provide landmark identification, and to simplify the data association problem. The proposed SLAM algorithm is implemented in real-time to further calibrate the ellipsoidal SLAM parameters, noise bounding, and to improve its overall accuracy. The augmented EKF/ellipsoidal SLAM algorithms are compared with the regular EKF/ellipsoidal SLAM methods and the merits of each algorithm is also discussed in the paper. The real-time experimental and simulation studies suggest that the adaptive augmented ellipsoidal SLAM is more accurate than the conventional EKF/ellipsoidal SLAMs.
Collapse
Affiliation(s)
- Elfituri S Lahemer
- Autonomous and Intelligent Systems Laboratory, School of Mechatronic Systems Engineering, Simon Fraser University, Surrey, BC, Canada.
| | - Ahmad Rad
- Autonomous and Intelligent Systems Laboratory, School of Mechatronic Systems Engineering, Simon Fraser University, Surrey, BC, Canada.
| |
Collapse
|
411
|
Advances in Nuclear Radiation Sensing: Enabling 3-D Gamma-Ray Vision. SENSORS 2019; 19:s19112541. [PMID: 31167360 PMCID: PMC6603681 DOI: 10.3390/s19112541] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/12/2019] [Revised: 05/28/2019] [Accepted: 06/01/2019] [Indexed: 11/17/2022]
Abstract
The enormous advances in sensing and data processing technologies in combination with recent developments in nuclear radiation detection and imaging enable unprecedented and “smarter” ways to detect, map, and visualize nuclear radiation. The recently developed concept of three-dimensional (3-D) Scene-data fusion allows us now to “see” nuclear radiation in three dimensions, in real time, and specific to radionuclides. It is based on a multi-sensor instrument that is able to map a local scene and to fuse the scene data with nuclear radiation data in 3-D while the instrument is freely moving through the scene. This new concept is agnostic of the deployment platform and the specific radiation detection or imaging modality. We have demonstrated this 3-D Scene-data fusion concept in a range of configurations in locations, such as the Fukushima Prefecture in Japan or Chernobyl in Ukraine on unmanned and manned aerial and ground-based platforms. It provides new means in the detection, mapping, and visualization of radiological and nuclear materials relevant for the safe and secure operation of nuclear and radiological facilities or in the response to accidental or intentional releases of radioactive materials where a timely, accurate, and effective assessment is critical. In addition, the ability to visualize nuclear radiation in 3-D and in real time provides new means in the communication with public and facilitates to overcome one of the major public concerns of not being able to “see” nuclear radiation.
Collapse
|
412
|
Gomez-Ojeda R, Moreno FA, Zuniga-Noel D, Scaramuzza D, Gonzalez-Jimenez J. PL-SLAM: A Stereo SLAM System Through the Combination of Points and Line Segments. IEEE T ROBOT 2019. [DOI: 10.1109/tro.2019.2899783] [Citation(s) in RCA: 146] [Impact Index Per Article: 29.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
413
|
Chen HY, Bell ZI, Deptula P, Dixon WE. A Switched Systems Approach to Path Following With Intermittent State Feedback. IEEE T ROBOT 2019. [DOI: 10.1109/tro.2019.2899269] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
414
|
Abstract
For autonomous driving, it is important to navigate in an unknown environment. In this paper, we propose a fully automated 2D simultaneous localization and mapping (SLAM) system based on lidar working in large-scale outdoor environments. To improve the accuracy and robustness of the scan matching module, an improved Correlative Scan Matching (CSM) algorithm is proposed. For efficient place recognition, we design an AdaBoost based loop closure detection algorithm which can efficiently reject false loop closures. For the SLAM back-end, we propose a light-weight graph optimization algorithm based on incremental smoothing and mapping (iSAM). The performance of our system is verified on various large-scale datasets including our real-world datasets and the KITTI odometry benchmark. Further comparisons to the state-of-the-art approaches indicate that our system is competitive with established techniques.
Collapse
|
415
|
Abstract
This paper focuses on loop-closure detection (LCD) for a visual simultaneous localization and mapping (SLAM) system. We present a strategy that combines a Bayes filter and features from a pre-trained convolution neural network (CNN) to perform LCD. Rather than using features from only one layer, we fuse features from multiple layers based on spatial pyramid pooling. A flexible Bayes model is then formulated to integrate the sequential information and similarities that are computed by features at different scales. The introduction of a penalty factor and bidirectional propagation enables our approach to handle complex trajectories. We present extensive experiments on challenging datasets, and we compare our approach to state-of-the-art methods, to evaluate it. The results show that our approach can ensure remarkable performance under severe condition changes and handle trajectories that have different characteristics. We also show the advantages of Bayes filters over sequence matching in the experiments, and we analyze our feature fusion strategy by visualizing the activations of the CNN.
Collapse
Affiliation(s)
- Qiang Liu
- School of Mechanical Engineering, Dalian University of Technology, Dalian 116024, Liaoning, P. R. China
| | - Fuhai Duan
- School of Mechanical Engineering, Dalian University of Technology, Dalian 116024, Liaoning, P. R. China
| |
Collapse
|
416
|
Moratuwage D, Adams M, Inostroza F. δ-Generalized Labeled Multi-Bernoulli Simultaneous Localization and Mapping with an Optimal Kernel-Based Particle Filtering Approach. SENSORS 2019; 19:s19102290. [PMID: 31108994 PMCID: PMC6567325 DOI: 10.3390/s19102290] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/16/2019] [Revised: 05/14/2019] [Accepted: 05/15/2019] [Indexed: 11/16/2022]
Abstract
Under realistic environmental conditions, heuristic-based data association and map management routines often result in divergent map and trajectory estimates in robotic Simultaneous Localization And Mapping (SLAM). To address these issues, SLAM solutions have been proposed based on the Random Finite Set (RFS) framework, which models the map and measurements such that the usual requirements of external data association routines and map management heuristics can be circumvented and realistic sensor detection uncertainty can be taken into account. Rao-Blackwellized particle filter (RBPF)-based RFS SLAM solutions have been demonstrated using the Probability Hypothesis Density (PHD) filter and subsequently the Labeled Multi-Bernoulli (LMB) filter. In multi-target tracking, the LMB filter, which was introduced as an efficient approximation to the computationally expensive δ -Generalized LMB ( δ -GLMB) filter, converts its representation of an LMB distribution to δ -GLMB form during the measurement update step. This not only results in a loss of information yielding inferior results (compared to the δ -GLMB filter) but also fails to take computational advantages in parallelized implementations possible with RBPF-based SLAM algorithms. Similar to state-of-the-art random vector-valued RBPF solutions such as FastSLAM and MH-FastSLAM, the performances of all RBPF-based SLAM algorithms based on the RFS framework also diverge from ground truth over time due to random sampling approaches, which only rely on control noise variance. Further, the methods lose particle diversity and diverge over time as a result of particle degeneracy. To alleviate this problem and further improve the quality of map estimates, a SLAM solution using an optimal kernel-based particle filter combined with an efficient variant of the δ -GLMB filter ( δ -GLMB-SLAM) is presented. The performance of the proposed δ -GLMB-SLAM algorithm, referred to as δ -GLMB-SLAM2.0, was demonstrated using simulated datasets and a section of the publicly available KITTI dataset. The results suggest that even with a limited number of particles, δ -GLMB-SLAM2.0 outperforms state-of-the-art RBPF-based RFS SLAM algorithms.
Collapse
Affiliation(s)
- Diluka Moratuwage
- Department of Electrical Engineering & Advanced Mining Technology Center Universidad de Chile, 837-0451 Santiago, Chile.
| | - Martin Adams
- Department of Electrical Engineering & Advanced Mining Technology Center Universidad de Chile, 837-0451 Santiago, Chile.
| | - Felipe Inostroza
- Department of Electrical Engineering Universidad de Chile, 837-0451 Santiago, Chile.
| |
Collapse
|
417
|
Intensity-Assisted ICP for Fast Registration of 2D-LIDAR. SENSORS 2019; 19:s19092124. [PMID: 31071958 PMCID: PMC6539496 DOI: 10.3390/s19092124] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/06/2019] [Revised: 04/12/2019] [Accepted: 04/29/2019] [Indexed: 11/16/2022]
Abstract
Iterative closest point (ICP) is a method commonly used to perform scan-matching and registration. To be a simple and robust algorithm, it is still computationally expensive, and it has been regarded as having a crucial challenge especially in a real-time application as used for the simultaneous localization and mapping (SLAM) problem. For these reasons, this paper presents a new method for the acceleration of ICP with an assisted intensity. Unlike the conventional ICP, this method is proposed to reduce the computational cost and avoid divergences. An initial transformation guess is computed with an assisted intensity for their relative rigid-body transformation. Moreover, a target function is proposed to determine the best initial transformation guess based on the statistic of their spatial distances and intensity residuals. Additionally, this method is also proposed to reduce the iteration number. The Anderson acceleration is utilized for increasing the iteration speed which has better ability than the Picard iteration procedure. The proposed algorithm is operated in real time with a single core central processing unit (CPU) thread. Hence, it is suitable for the robot which has limited computation resources. To validate the novelty, this proposed method is evaluated on the SEMANTIC3D.NET benchmark dataset. According to comparative results, the proposed method is declared as having better accuracy and robustness than the conventional ICP methods.
Collapse
|
418
|
GNSS/INS/LiDAR-SLAM Integrated Navigation System Based on Graph Optimization. REMOTE SENSING 2019. [DOI: 10.3390/rs11091009] [Citation(s) in RCA: 44] [Impact Index Per Article: 8.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
A Global Navigation Satellite System (GNSS)/Inertial Navigation System (INS)/Light Detection and Ranging (LiDAR)-Simultaneous Localization and Mapping (SLAM) integrated navigation system based on graph optimization is proposed and implemented in this paper. The navigation results are obtained by the information fusion of the GNSS position, Inertial Measurement Unit (IMU) preintegration result and the relative pose from the 3D probability map matching with graph optimizing. The sliding window method was adopted to ensure that the computational load of the graph optimization does not increase with time. Land vehicle tests were conducted, and the results show that the proposed GNSS/INS/LiDAR-SLAM integrated navigation system can effectively improve the navigation positioning accuracy compared to GNSS/INS and other current GNSS/INS/LiDAR methods. During the simulation of one-minute periods of GNSS outages, compared to the GNSS/INS integrated navigation system, the root mean square (RMS) of the position errors in the North and East directions of the proposed navigation system are reduced by approximately 82.2% and 79.6%, respectively, and the position error in the vertical direction and attitude errors are equivalent. Compared to the benchmark method of GNSS/INS/LiDAR-Google Cartographer, the RMS of the position errors in the North, East and vertical directions decrease by approximately 66.2%, 63.1% and 75.1%, respectively, and the RMS of the roll, pitch and yaw errors are reduced by approximately 89.5%, 92.9% and 88.5%, respectively. Furthermore, the relative position error during the GNSS outage periods is reduced to 0.26% of the travel distance for the proposed method. Therefore, the GNSS/INS/LiDAR-SLAM integrated navigation system proposed in this paper can effectively fuse the information of GNSS, IMU and LiDAR and can significantly mitigate the navigation error, especially for cases of GNSS signal attenuation or interruption.
Collapse
|
419
|
Abstract
For autonomous driving, it is important to obtain precise and high-frequency localization information. This paper proposes a novel method in which the Inertial Measurement Unit (IMU), wheel encoder, and lidar odometry are utilized together to estimate the ego-motion of an unmanned ground vehicle. The IMU is fused with the wheel encoder to obtain the motion prior, and it is involved in three levels of the lidar odometry: Firstly, we use the IMU information to rectify the intra-frame distortion of the lidar scan, which is caused by the vehicle’s own movement; secondly, the IMU provides a better initial guess for the lidar odometry; and thirdly, the IMU is fused with the lidar odometry in an Extended Kalman filter framework. In addition, an efficient method for hand–eye calibration between the IMU and the lidar is proposed. To evaluate the performance of our method, extensive experiments are performed and our system can output stable, accurate, and high-frequency localization results in diverse environment without any prior information.
Collapse
|
420
|
Garg S, Suenderhauf N, Milford M. Semantic–geometric visual place recognition: a new perspective for reconciling opposing views. Int J Rob Res 2019. [DOI: 10.1177/0278364919839761] [Citation(s) in RCA: 28] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Affiliation(s)
- Sourav Garg
- Australian Centre for Robotic Vision, Queensland University of Technology, Brisbane, Australia
| | - Niko Suenderhauf
- Australian Centre for Robotic Vision, Queensland University of Technology, Brisbane, Australia
| | - Michael Milford
- Australian Centre for Robotic Vision, Queensland University of Technology, Brisbane, Australia
| |
Collapse
|
421
|
Estimating Pavement Roughness by Fusing Color and Depth Data Obtained from an Inexpensive RGB-D Sensor. SENSORS 2019; 19:s19071655. [PMID: 30959936 PMCID: PMC6479490 DOI: 10.3390/s19071655] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/13/2019] [Revised: 03/26/2019] [Accepted: 03/27/2019] [Indexed: 11/17/2022]
Abstract
Measuring pavement roughness and detecting pavement surface defects are two of the most important tasks in pavement management. While existing pavement roughness measurement approaches are expensive, the primary aim of this paper is to use a cost-effective and sufficiently accurate RGB-D sensor to estimate the pavement roughness in the outdoor environment. An algorithm is proposed to process the RGB-D data and autonomously quantify the road roughness. To this end, the RGB-D sensor is calibrated and primary data for estimating the pavement roughness are collected. The collected depth frames and RGB images are registered to create the 3D road surfaces. We found that there is a significant correlation between the estimated International Roughness Index (IRI) using the RGB-D sensor and the manual measured IRI using rod and level. By considering the Power Spectral Density (PSD) analysis and the repeatability of measurement, the results show that the proposed solution can accurately estimate the different pavement roughness.
Collapse
|
422
|
Eckenhoff K, Geneva P, Huang G. Closed-form preintegration methods for graph-based visual–inertial navigation. Int J Rob Res 2019. [DOI: 10.1177/0278364919835021] [Citation(s) in RCA: 31] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
In this paper, we propose a new analytical preintegration theory for graph-based sensor fusion with an inertial measurement unit (IMU) and a camera (or other aiding sensors). Rather than using discrete sampling of the measurement dynamics as in current methods, we derive the closed-form solutions to the preintegration equations, yielding improved accuracy in state estimation. We advocate two new different inertial models for preintegration: (i) the model that assumes piecewise constant measurements; and (ii) the model that assumes piecewise constant local true acceleration. Through extensive Monte Carlo simulations, we show the effect that the choice of preintegration model has on estimation performance. To validate the proposed preintegration theory, we develop both direct and indirect visual–inertial navigation systems (VINSs) that leverage our preintegration. In the first, within a tightly coupled, sliding-window optimization framework, we jointly estimate the features in the window and the IMU states while performing marginalization to bound the computational cost. In the second, we loosely couple the IMU preintegration with a direct image alignment that estimates relative camera motion by minimizing the photometric errors (i.e., image intensity difference), allowing for efficient and informative loop closures. Both systems are extensively validated in real-world experiments and are shown to offer competitive performance to state-of-the-art methods.
Collapse
Affiliation(s)
- Kevin Eckenhoff
- Department of Mechanical Engineering, and Computer and Information Sciences, University of Delaware, Newark, DE, USA
| | - Patrick Geneva
- Department of Mechanical Engineering, and Computer and Information Sciences, University of Delaware, Newark, DE, USA
| | - Guoquan Huang
- Department of Mechanical Engineering, and Computer and Information Sciences, University of Delaware, Newark, DE, USA
| |
Collapse
|
423
|
Gostar AK, Fu C, Chuah W, Hossain MI, Tennakoon R, Bab-Hadiashar A, Hoseinnezhad R. State Transition for Statistical SLAM Using Planar Features in 3D Point Clouds. SENSORS 2019; 19:s19071614. [PMID: 30987259 PMCID: PMC6479366 DOI: 10.3390/s19071614] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/11/2019] [Revised: 03/23/2019] [Accepted: 04/01/2019] [Indexed: 12/02/2022]
Abstract
There is a large body of literature on solving the SLAM problem for various autonomous vehicle applications. A substantial part of the solutions is formulated based on using statistical (mainly Bayesian) filters such as Kalman filter and its extended version. In such solutions, the measurements are commonly some point features or detections collected by the sensor(s) on board the autonomous vehicle. With the increasing utilization of scanners with common autonomous cars, and availability of 3D point clouds in real-time and at fast rates, it is now possible to use more sophisticated features extracted from the point clouds for filtering. This paper presents the idea of using planar features with multi-object Bayesian filters for SLAM. With Bayesian filters, the first step is prediction, where the object states are propagated to the next time based on a stochastic transition model. We first present how such a transition model can be developed, and then propose a solution for state prediction. In the simulation studies, using a dataset of measurements acquired from real vehicle sensors, we apply the proposed model to predict the next planar features and vehicle states. The results show reasonable accuracy and efficiency for statistical filtering-based SLAM applications.
Collapse
Affiliation(s)
| | - Chunyun Fu
- State Key Laboratory of Mechanical Transmissions, School of Automotive Engineering, Chongqing University, Chongqing 400044, China.
| | - Weiqin Chuah
- School of Engineering, RMIT University, Melbourne VIC 3001, Australia.
| | | | - Ruwan Tennakoon
- School of Engineering, RMIT University, Melbourne VIC 3001, Australia.
| | | | - Reza Hoseinnezhad
- School of Engineering, RMIT University, Melbourne VIC 3001, Australia.
| |
Collapse
|
424
|
Liu L, Zhang T, Leighton B, Zhao L, Huang S, Dissanayake G. Robust Global Structure From Motion Pipeline With Parallax on Manifold Bundle Adjustment and Initialization. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2900756] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
|
425
|
Chiang HTL, Faust A, Fiser M, Francis A. Learning Navigation Behaviors End-to-End With AutoRL. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2899918] [Citation(s) in RCA: 80] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
426
|
Eckenhoff K, Yang Y, Geneva P, Huang G. Tightly-Coupled Visual-Inertial Localization and 3-D Rigid-Body Target Tracking. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2896472] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
427
|
Tsintotas KA, Bampis L, Gasteratos A. Probabilistic Appearance-Based Place Recognition Through Bag of Tracked Words. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2897151] [Citation(s) in RCA: 28] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
428
|
Yang K, Fang W, Zhao Y, Deng N. Iteratively Reweighted Midpoint Method for Fast Multiple View Triangulation. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2893022] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
429
|
Nardi F, Della Corte B, Grisetti G. Unified representation and registration of heterogeneous sets of geometric primitives. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2891989] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
430
|
Marge M, Rudnicky AI. Miscommunication Detection and Recovery in Situated Human–Robot Dialogue. ACM T INTERACT INTEL 2019. [DOI: 10.1145/3237189] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
Abstract
Even without speech recognition errors, robots may face difficulties interpreting natural-language instructions. We present a method for robustly handling miscommunication between people and robots in task-oriented spoken dialogue. This capability is implemented in TeamTalk, a conversational interface to robots that supports detection and recovery from the
situated grounding problems
of referential ambiguity and impossible actions. We introduce a representation that detects these problems and a nearest-neighbor learning algorithm that selects recovery strategies for a virtual robot. When the robot encounters a grounding problem, it looks back on its interaction history to consider how it resolved similar situations. The learning method is trained initially on crowdsourced data but is then supplemented by interactions from a longitudinal user study in which six participants performed navigation tasks with the robot. We compare results collected using a general model to user-specific models and find that user-specific models perform best on measures of dialogue efficiency, while the general model yields the highest agreement with human judges. Our overall contribution is a novel approach to detecting and recovering from miscommunication in dialogue by including situated context, namely, information from a robot’s path planner and surroundings.
Collapse
|
431
|
A New Method of Simultaneous Localization and Mapping for Mobile Robots Using Acoustic Landmarks. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9071352] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The simultaneous localization and mapping (SLAM) problem for mobile robots has always been a hotspot in the field of robotics. Simultaneous localization and mapping for robots using visual sensors and laser radar is easily affected by the field of view and ground conditions. According to the problems of traditional sensors applied in SLAM, this paper presents a novel method to perform SLAM using acoustic signals. This method enables robots equipped with sound sources, moving within a working environment and interacting with microphones of interest, to locate itself and map the objects simultaneously. In our case, a method of microphone localization based on a sound source array is proposed, and it was applied as a pre-processing step to the SLAM procedure. A microphone capable of receiving sound signals can be directly used as a feature landmark of a robot observation model without feature extraction. Meanwhile, to eliminate the random error caused by hardware equipment, a sound settled in the middle of two microphones was applied as a calibration sound source to determine the value of the random error. Simulations and realistic experimental results demonstrate the feasibility and effectiveness of the proposed method.
Collapse
|
432
|
Chen G, Chen J, Lienen M, Conradt J, Röhrbein F, Knoll AC. FLGR: Fixed Length Gists Representation Learning for RNN-HMM Hybrid-Based Neuromorphic Continuous Gesture Recognition. Front Neurosci 2019; 13:73. [PMID: 30809114 PMCID: PMC6380225 DOI: 10.3389/fnins.2019.00073] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2018] [Accepted: 01/23/2019] [Indexed: 11/18/2022] Open
Abstract
A neuromorphic vision sensors is a novel passive sensing modality and frameless sensors with several advantages over conventional cameras. Frame-based cameras have an average frame-rate of 30 fps, causing motion blur when capturing fast motion, e.g., hand gesture. Rather than wastefully sending entire images at a fixed frame rate, neuromorphic vision sensors only transmit the local pixel-level changes induced by the movement in a scene when they occur. This leads to advantageous characteristics, including low energy consumption, high dynamic range, a sparse event stream and low response latency. In this study, a novel representation learning method was proposed: Fixed Length Gists Representation (FLGR) learning for event-based gesture recognition. Previous methods accumulate events into video frames in a time duration (e.g., 30 ms) to make the accumulated image-level representation. However, the accumulated-frame-based representation waives the friendly event-driven paradigm of neuromorphic vision sensor. New representation are urgently needed to fill the gap in non-accumulated-frame-based representation and exploit the further capabilities of neuromorphic vision. The proposed FLGR is a sequence learned from mixture density autoencoder and preserves the nature of event-based data better. FLGR has a data format of fixed length, and it is easy to feed to sequence classifier. Moreover, an RNN-HMM hybrid was proposed to address the continuous gesture recognition problem. Recurrent neural network (RNN) was applied for FLGR sequence classification while hidden Markov model (HMM) is employed for localizing the candidate gesture and improving the result in a continuous sequence. A neuromorphic continuous hand gestures dataset (Neuro ConGD Dataset) was developed with 17 hand gestures classes for the community of the neuromorphic research. Hopefully, FLGR can inspire the study on the event-based highly efficient, high-speed, and high-dynamic-range sequence classification tasks.
Collapse
Affiliation(s)
- Guang Chen
- College of Automotive Engineering, Tongji University, Shanghai, China.,Chair of Robotics, Artificial Intelligence and Real-time Systems, Technische Universität München, Munich, Germany
| | - Jieneng Chen
- College of Electronics and Information Engineering, Tongji University, Shanghai, China
| | - Marten Lienen
- Chair of Robotics, Artificial Intelligence and Real-time Systems, Technische Universität München, Munich, Germany
| | - Jörg Conradt
- Department of Computational Science and Technology, KTH Royal Institute of Technology, Stockholm, Sweden
| | - Florian Röhrbein
- Chair of Robotics, Artificial Intelligence and Real-time Systems, Technische Universität München, Munich, Germany
| | - Alois C Knoll
- Chair of Robotics, Artificial Intelligence and Real-time Systems, Technische Universität München, Munich, Germany
| |
Collapse
|
433
|
Abstract
The state-of-the-art visual simultaneous localization and mapping (V-SLAM) systems have high accuracy localization capabilities and impressive mapping effects. However, most of these systems assume that the operating environment is static, thereby limiting their application in the real dynamic world. In this paper, by fusing the information of an RGB-D camera and two encoders that are mounted on a differential-drive robot, we aim to estimate the motion of the robot and construct a static background OctoMap in both dynamic and static environments. A tightly coupled feature-based method is proposed to fuse the two types of information based on the optimization. Dynamic pixels occupied by dynamic objects are detected and culled to cope with dynamic environments. The ability to identify the dynamic pixels on both predefined and undefined dynamic objects is available, which is attributed to the combination of the CPU-based object detection method and a multiview constraint-based approach. We first construct local sub-OctoMaps by using the keyframes and then fuse the sub-OctoMaps into a full OctoMap. This submap-based approach gives the OctoMap the ability to deform, and significantly reduces the map updating time and memory costs. We evaluated the proposed system in various dynamic and static scenes. The results show that our system possesses competitive pose accuracy and high robustness, as well as the ability to construct a clean static OctoMap in dynamic scenes.
Collapse
|
434
|
Real-Time Monocular Visual Odometry for Turbid and Dynamic Underwater Environments. SENSORS 2019; 19:s19030687. [PMID: 30743993 PMCID: PMC6386985 DOI: 10.3390/s19030687] [Citation(s) in RCA: 28] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/26/2018] [Revised: 02/01/2019] [Accepted: 02/06/2019] [Indexed: 12/05/2022]
Abstract
In the context of underwater robotics, the visual degradation induced by the medium properties make difficult the exclusive use of cameras for localization purpose. Hence, many underwater localization methods are based on expensive navigation sensors associated with acoustic positioning. On the other hand, pure visual localization methods have shown great potential in underwater localization but the challenging conditions, such as the presence of turbidity and dynamism, remain complex to tackle. In this paper, we propose a new visual odometry method designed to be robust to these visual perturbations. The proposed algorithm has been assessed on both simulated and real underwater datasets and outperforms state-of-the-art terrestrial visual SLAM methods under many of the most challenging conditions. The main application of this work is the localization of Remotely Operated Vehicles used for underwater archaeological missions, but the developed system can be used in any other applications as long as visual information is available.
Collapse
|
435
|
|
436
|
Abstract
The timely and efficient generation of detailed damage maps is of fundamental importance following disaster events to speed up first responders’ (FR) rescue activities and help trapped victims. Several works dealing with the automated detection of building damages have been published in the last decade. The increasingly widespread availability of inexpensive UAV platforms has also driven their recent adoption for rescue operations (i.e., search and rescue). Their deployment, however, remains largely limited to visual image inspection by skilled operators, limiting their applicability in time-constrained real conditions. This paper proposes a new solution to autonomously map building damages with a commercial UAV in near real-time. The solution integrates different components that allow the live streaming of the images on a laptop and their processing on the fly. Advanced photogrammetric techniques and deep learning algorithms are combined to deliver a true-orthophoto showing the position of building damages, which are already processed by the time the UAV returns to base. These algorithms have been customized to deliver fast results, fulfilling the near real-time requirements. The complete solution has been tested in different conditions, and received positive feedback by the FR involved in the EU funded project INACHUS. Two realistic pilot tests are described in the paper. The achieved results show the great potential of the presented approach, how close the proposed solution is to FR’ expectations, and where more work is still needed.
Collapse
|
437
|
Freda L, Gianni M, Pirri F, Gawel A, Dubé R, Siegwart R, Cadena C. 3D multi-robot patrolling with a two-level coordination strategy. Auton Robots 2019. [DOI: 10.1007/s10514-018-09822-3] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
438
|
Muraccini M, Mangia AL, Lannocca M, Cappello A. Magnetometer Calibration and Field Mapping through Thin Plate Splines. SENSORS 2019; 19:s19020280. [PMID: 30641986 PMCID: PMC6359173 DOI: 10.3390/s19020280] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/14/2018] [Revised: 12/21/2018] [Accepted: 01/07/2019] [Indexed: 11/16/2022]
Abstract
While the undisturbed Earth’s magnetic field represents a fundamental information source for orientation purposes, magnetic distortions have been mostly considered as a source of error. However, when distortions are temporally stable and spatially distinctive, they could provide a unique magnetic landscape that can be used in different applications, from indoor localization to sensor fusion algorithms for attitude estimation. The main purpose of this work, therefore, is to present a method to characterize the 3D magnetic vector in every point of the measurement volume. The possibility of describing the 3D magnetic field map through Thin Plate Splines (TPS) interpolation is investigated and demonstrated. An algorithm for the simultaneous estimation of the parameters related to magnetometer calibration and those describing the magnetic map, is proposed and tested on both simulated and real data. Results demonstrate that an accurate description of the local magnetic field using TPS interpolation is possible. The proposed procedure leads to errors in the estimation of the local magnetic direction with a standard deviation lower than 1 degree. Magnetometer calibration and magnetic field mapping could be integrated into different algorithms, for example to improve attitude estimation in highly distorted environments or as an aid to indoor localization.
Collapse
Affiliation(s)
- Marco Muraccini
- Department of Electrical, Electronic and Information Engineering, University of Bologna, Viale del Risorgimento, 2, 40136 Bologna, Italy.
| | - Anna Lisa Mangia
- Department of Electrical, Electronic and Information Engineering, University of Bologna, Viale del Risorgimento, 2, 40136 Bologna, Italy.
| | - Maurizio Lannocca
- Department of Electrical, Electronic and Information Engineering, University of Bologna, Viale del Risorgimento, 2, 40136 Bologna, Italy.
| | - Angelo Cappello
- Department of Electrical, Electronic and Information Engineering, University of Bologna, Viale del Risorgimento, 2, 40136 Bologna, Italy.
| |
Collapse
|
439
|
Zhou X, Gao Y, Guan L. Towards Goal-Directed Navigation Through Combining Learning Based Global and Local Planners. SENSORS (BASEL, SWITZERLAND) 2019; 19:E176. [PMID: 30621314 PMCID: PMC6339171 DOI: 10.3390/s19010176] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/13/2018] [Revised: 12/13/2018] [Accepted: 12/25/2018] [Indexed: 11/23/2022]
Abstract
Robot navigation is a fundamental problem in robotics and various approaches have been developed to cope with this problem. Despite the great success of previous approaches, learning-based methods are receiving growing interest in the research community. They have shown great efficiency in solving navigation tasks and offer considerable promise to build intelligent navigation systems. This paper presents a goal-directed robot navigation system that integrates global planning based on goal-directed end-to-end learning and local planning based on reinforcement learning (RL). The proposed system aims to navigate the robot to desired goal positions while also being adaptive to changes in the environment. The global planner is trained to imitate an expert's navigation between different positions by goal-directed end-to-end learning, where both the goal representations and local observations are incorporated to generate actions. However, it is trained in a supervised fashion and is weak in dealing with changes in the environment. To solve this problem, a local planner based on deep reinforcement learning (DRL) is designed. The local planner is first implemented in a simulator and then transferred to the real world. It works complementarily to deal with situations that have not been met during training the global planner and is able to generalize over different situations. The experimental results on a robot platform demonstrate the effectiveness of the proposed navigation system.
Collapse
Affiliation(s)
- Xiaomao Zhou
- College of Automation, Harbin Engineering University, Harbin 150001, China.
| | - Yanbin Gao
- College of Automation, Harbin Engineering University, Harbin 150001, China.
| | - Lianwu Guan
- College of Automation, Harbin Engineering University, Harbin 150001, China.
| |
Collapse
|
440
|
Visual Semantic Landmark-Based Robust Mapping and Localization for Autonomous Indoor Parking. SENSORS 2019; 19:s19010161. [PMID: 30621195 PMCID: PMC6338888 DOI: 10.3390/s19010161] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/09/2018] [Revised: 12/28/2018] [Accepted: 12/31/2018] [Indexed: 11/17/2022]
Abstract
Autonomous parking in an indoor parking lot without human intervention is one of the most demanded and challenging tasks of autonomous driving systems. The key to this task is precise real-time indoor localization. However, state-of-the-art low-level visual feature-based simultaneous localization and mapping systems (VSLAM) suffer in monotonous or texture-less scenes and under poor illumination or dynamic conditions. Additionally, low-level feature-based mapping results are hard for human beings to use directly. In this paper, we propose a semantic landmark-based robust VSLAM for real-time localization of autonomous vehicles in indoor parking lots. The parking slots are extracted as meaningful landmarks and enriched with confidence levels. We then propose a robust optimization framework to solve the aliasing problem of semantic landmarks by dynamically eliminating suboptimal constraints in the pose graph and correcting erroneous parking slots associations. As a result, a semantic map of the parking lot, which can be used by both autonomous driving systems and human beings, is established automatically and robustly. We evaluated the real-time localization performance using multiple autonomous vehicles, and an repeatability of 0.3 m track tracing was achieved at a 10 kph of autonomous driving.
Collapse
|
441
|
A Multi-User Personal Indoor Localization System Employing Graph-Based Optimization. SENSORS 2019; 19:s19010157. [PMID: 30621181 PMCID: PMC6338911 DOI: 10.3390/s19010157] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/06/2018] [Revised: 12/30/2018] [Accepted: 12/31/2018] [Indexed: 11/23/2022]
Abstract
Personal indoor localization with smartphones is a well-researched area, with a number of approaches solving the problem separately for individual users. Most commonly, a particle filter is used to fuse information from dead reckoning and WiFi or Bluetooth adapters to provide an accurate location of the person holding a smartphone. Unfortunately, the existing solutions largely ignore the gains that emerge when a single localization system estimates locations of multiple users in the same environment. Approaches based on filtration maintain only estimates of the current poses of the users, marginalizing the historical data. Therefore, it is difficult to fuse data from multiple individual trajectories that are usually not perfectly synchronized in time. We propose a system that fuses the information from WiFi and dead reckoning employing the graph-based optimization, which is widely applied in robotics. The presented system can be used for localization of a single user, but the improvement is especially visible when this approach is extended to a multi-user scenario. The article presents a number of experiments performed with a smartphone inside an office building. These experiments demonstrate that graph-based optimization can be used as an efficient fusion mechanism to obtain accurate trajectory estimates both in the case of a single user and in a multi-user indoor localization system. The code of our system together with recorded dataset will be made available when the paper gets published.
Collapse
|
442
|
Tang J, Ericson L, Folkesson J, Jensfelt P. GCNv2: Efficient Correspondence Prediction for Real-Time SLAM. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2927954] [Citation(s) in RCA: 36] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
443
|
Bernreiter L, Gawel AR, Sommer H, Nieto J, Siegwart R, Cadena Lerma C. Multiple Hypothesis Semantic Mapping for Robust Data Association. IEEE Robot Autom Lett 2019. [DOI: 10.1109/lra.2019.2925756] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
444
|
A Robust Registration Method for Autonomous Driving Pose Estimation in Urban Dynamic Environment Using LiDAR. ELECTRONICS 2019. [DOI: 10.3390/electronics8010043] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The registration of point clouds in urban environments faces problems such as dynamic vehicles and pedestrians, changeable road environments, and GPS inaccuracies. The state-of-the-art methodologies have usually combined the dynamic object tracking and/or static feature extraction data into a point cloud towards the solution of these problems. However, there is the occurrence of minor initial position errors due to these methodologies. In this paper, the authors propose a fast and robust registration method that exhibits no need for the detection of any dynamic and/or static objects. This proposed methodology may be able to adapt to higher initial errors. The initial steps of this methodology involved the optimization of the object segmentation under the application of a series of constraints. Based on this algorithm, a novel multi-layer nested RANSAC algorithmic framework is proposed to iteratively update the registration results. The robustness and efficiency of this algorithm is demonstrated on several high dynamic scenes of both short and long time intervals with varying initial offsets. A LiDAR odometry experiment was performed on the KITTI data set and our extracted urban data-set with a high dynamic urban road, and the average of the horizontal position errors was compared to the distance traveled that resulted in 0.45% and 0.55% respectively.
Collapse
|
445
|
Abstract
Trainable visual navigation systems based on deep learning demonstrate potential for robustness of onboard camera parameters and challenging environment. However, a deep model requires substantial computational resources and large labelled training sets for successful training. Implementation of the autonomous navigation and training-based fast adaptation to the new environment for a compact drone is a complicated task. The article describes an original model and training algorithms adapted to the limited volume of labelled training set and constrained computational resource. This model consists of a convolutional neural network for visual feature extraction, extreme-learning machine for estimating the position displacement and boosted information-extreme classifier for obstacle prediction. To perform unsupervised training of the convolution filters with a growing sparse-coding neural gas algorithm, supervised learning algorithms to construct the decision rules with simulated annealing search algorithm used for finetuning are proposed. The use of complex criterion for parameter optimization of the feature extractor model is considered. The resulting approach performs better trajectory reconstruction than the well-known ORB-SLAM. In particular, for sequence 7 from the KITTI dataset, the translation error is reduced by nearly 65.6% under the frame rate 10 frame per second. Besides, testing on the independent TUM sequence shot outdoors produces a translation error not exceeding 6% and a rotation error not exceeding 3.68 degrees per 100 m. Testing was carried out on the Raspberry Pi 3+ single-board computer.
Collapse
|
446
|
Abstract
SUMMARYA Q-learning approach is often used for navigation in static environments where state space is easy to define. In this paper, a new Q-learning approach is proposed for navigation in dynamic environments by imitating human reasoning. As a model-free method, a Q-learning method does not require the environmental model in advance. The state space and the reward function in the proposed approach are defined according to human perception and evaluation, respectively. Specifically, approximate regions instead of accurate measurements are used to define states. Moreover, due to the limitation of robot dynamics, actions for each state are calculated by introducing a dynamic window that takes robot dynamics into account. The conducted tests show that the obstacle avoidance rate of the proposed approach can reach 90.5% after training, and the robot can always operate below the dynamics limitation.
Collapse
|
447
|
Schuster MJ, Schmid K, Brand C, Beetz M. Distributed stereo vision-based 6D localization and mapping for multi-robot teams. J FIELD ROBOT 2018. [DOI: 10.1002/rob.21812] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Affiliation(s)
- Martin J. Schuster
- Department of Perception and Cognition, Robotics and Mechatronics Center (RMC); German Aerospace Center (DLR); Weßling Germany
| | | | - Christoph Brand
- Department of Perception and Cognition, Robotics and Mechatronics Center (RMC); German Aerospace Center (DLR); Weßling Germany
| | - Michael Beetz
- Institute for Artificial Intelligence and Center for Computing Technologies (TZI); Faculty of Computer Science, University Bremen; Bremen Germany
| |
Collapse
|
448
|
Abstract
SUMMARYA robot should be able to estimate an accurate and dense 3D model of its environment (a map), along with its pose relative to it, all of it in real time, in order to be able to navigate autonomously without collisions.As the robot moves from its starting position and the estimated map grows, the computational and memory footprint of a dense 3D map increases and might exceed the robot capabilities in a short time. However, a global map is still needed to maintain its consistency and plan for distant goals, possibly out of the robot field of view.In this work, we address such problem by proposing a real-time stereo mapping pipeline, feasible for standard CPUs, which is locally dense and globally sparse and accurate. Our algorithm is based on a graph relating poses and salient visual points, in order to maintain a long-term accuracy with a small cost. Within such framework, we propose an efficient dense fusion of several stereo depths in the locality of the current robot pose.We evaluate the performance and the accuracy of our algorithm in the public datasets of Tsukuba and KITTI, and demonstrate that it outperforms single-view stereo depth. We release the code as open-source, in order to facilitate the system use and comparisons.
Collapse
|
449
|
Kunze L, Hawes N, Duckett T, Hanheide M, Krajnik T. Artificial Intelligence for Long-Term Robot Autonomy: A Survey. IEEE Robot Autom Lett 2018. [DOI: 10.1109/lra.2018.2860628] [Citation(s) in RCA: 74] [Impact Index Per Article: 12.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
450
|
Li R, Wang S, Gu D. Ongoing Evolution of Visual SLAM from Geometry to Deep Learning: Challenges and Opportunities. Cognit Comput 2018. [DOI: 10.1007/s12559-018-9591-8] [Citation(s) in RCA: 29] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|