1
|
Lin Z, Tian Z, Zhang Q, Zhuang H, Lan J. Enhanced Visual SLAM for Collision-Free Driving with Lightweight Autonomous Cars. SENSORS (BASEL, SWITZERLAND) 2024; 24:6258. [PMID: 39409298 PMCID: PMC11478337 DOI: 10.3390/s24196258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Revised: 09/23/2024] [Accepted: 09/25/2024] [Indexed: 10/20/2024]
Abstract
The paper presents a vision-based obstacle avoidance strategy for lightweight self-driving cars that can be run on a CPU-only device using a single RGB-D camera. The method consists of two steps: visual perception and path planning. The visual perception part uses ORBSLAM3 enhanced with optical flow to estimate the car's poses and extract rich texture information from the scene. In the path planning phase, the proposed method employs a method combining a control Lyapunov function and control barrier function in the form of a quadratic program (CLF-CBF-QP) together with an obstacle shape reconstruction process (SRP) to plan safe and stable trajectories. To validate the performance and robustness of the proposed method, simulation experiments were conducted with a car in various complex indoor environments using the Gazebo simulation environment. The proposed method can effectively avoid obstacles in the scenes. The proposed algorithm outperforms benchmark algorithms in achieving more stable and shorter trajectories across multiple simulated scenes.
Collapse
Affiliation(s)
- Zhihao Lin
- James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK; (Z.L.); (Z.T.)
| | - Zhen Tian
- James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK; (Z.L.); (Z.T.)
| | - Qi Zhang
- Faculty of Science, University of Amsterdam, Science Park 904, 1098 XH Amsterdam, The Netherlands;
| | - Hanyang Zhuang
- University of Michigan-Shanghai Jiao Tong University Joint Institute, Shanghai Jiao Tong University, Shanghai 200240, China;
| | - Jianglin Lan
- James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK; (Z.L.); (Z.T.)
| |
Collapse
|
2
|
Dou H, Wang Z, Wang C, Zhao X. Immediate Pose Recovery Method for Untracked Frames in Feature-Based SLAM. SENSORS (BASEL, SWITZERLAND) 2024; 24:835. [PMID: 38339551 PMCID: PMC10857547 DOI: 10.3390/s24030835] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/30/2023] [Revised: 01/22/2024] [Accepted: 01/25/2024] [Indexed: 02/12/2024]
Abstract
In challenging environments, feature-based visual SLAM encounters frequent failures in frame tracking, introducing unknown poses to robotic applications. This paper introduces an immediate approach for recovering untracked camera poses. Through the retrieval of key information from elapsed untracked frames, lost poses are efficiently restored with a short time consumption. Taking account of reconstructed poses and map points during local optimizing, a denser local map is constructed around ambiguous frames to enhance the further SLAM procedure. The proposed method is implemented in a SLAM system, and monocular experiments are conducted on datasets. The experimental results demonstrate that our method can reconstruct the untracked frames in nearly real time, effectively complementing missing segments of the trajectory. Concurrently, the accuracy and robustness for subsequent tracking are improved through the integration of recovered poses and map points.
Collapse
Affiliation(s)
| | | | - Changhong Wang
- Space Control and Inertial Technology Research Center, School of Astronautics, Harbin Institute of Technology, Harbin 150001, China
| | | |
Collapse
|
3
|
Zhang X, Yu H, Zhuang Y. A robust RGB‐D visual odometry with moving object detection in dynamic indoor scenes. IET CYBER-SYSTEMS AND ROBOTICS 2023. [DOI: 10.1049/csy2.12079] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2023] Open
Affiliation(s)
- Xianglong Zhang
- College of Mechanical and Electronic Engineering Dalian Minzu University Dalian China
| | - Haiyang Yu
- College of Mechanical and Electronic Engineering Dalian Minzu University Dalian China
| | - Yan Zhuang
- School of Control Science and Engineering Dalian University of Technology Dalian China
| |
Collapse
|
4
|
Tourani A, Bavle H, Sanchez-Lopez JL, Voos H. Visual SLAM: What Are the Current Trends and What to Expect? SENSORS (BASEL, SWITZERLAND) 2022; 22:9297. [PMID: 36501998 PMCID: PMC9735432 DOI: 10.3390/s22239297] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 11/21/2022] [Accepted: 11/25/2022] [Indexed: 06/17/2023]
Abstract
In recent years, Simultaneous Localization and Mapping (SLAM) systems have shown significant performance, accuracy, and efficiency gain. In this regard, Visual Simultaneous Localization and Mapping (VSLAM) methods refer to the SLAM approaches that employ cameras for pose estimation and map reconstruction and are preferred over Light Detection And Ranging (LiDAR)-based methods due to their lighter weight, lower acquisition costs, and richer environment representation. Hence, several VSLAM approaches have evolved using different camera types (e.g., monocular or stereo), and have been tested on various datasets (e.g., Technische Universität München (TUM) RGB-D or European Robotics Challenge (EuRoC)) and in different conditions (i.e., indoors and outdoors), and employ multiple methodologies to have a better understanding of their surroundings. The mentioned variations have made this topic popular for researchers and have resulted in various methods. In this regard, the primary intent of this paper is to assimilate the wide range of works in VSLAM and present their recent advances, along with discussing the existing challenges and trends. This survey is worthwhile to give a big picture of the current focuses in robotics and VSLAM fields based on the concentrated resolutions and objectives of the state-of-the-art. This paper provides an in-depth literature survey of fifty impactful articles published in the VSLAMs domain. The mentioned manuscripts have been classified by different characteristics, including the novelty domain, objectives, employed algorithms, and semantic level. The paper also discusses the current trends and contemporary directions of VSLAM techniques that may help researchers investigate them.
Collapse
Affiliation(s)
- Ali Tourani
- Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg
| | - Hriday Bavle
- Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg
| | - Jose Luis Sanchez-Lopez
- Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg
| | - Holger Voos
- Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg
- Faculty of Science, Technology and Medicine (FSTM), Department of Engineering, University of Luxembourg, 1359 Luxembourg, Luxembourg
| |
Collapse
|
5
|
A Novel Method for Distinguishing Indoor Dynamic and Static Semantic Objects Based on Deep Learning and Space Constraints in Visual-inertial SLAM. J INTELL ROBOT SYST 2022. [DOI: 10.1007/s10846-022-01730-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
6
|
An improved adaptive ORB-SLAM method for monocular vision robot under dynamic environments. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-022-01627-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
7
|
Yang K, Zhang W, Li C, Wang X. Accurate Location in Dynamic Traffic Environment Using Semantic Information and Probabilistic Data Association. SENSORS 2022; 22:s22135042. [PMID: 35808536 PMCID: PMC9269809 DOI: 10.3390/s22135042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Revised: 06/10/2022] [Accepted: 06/29/2022] [Indexed: 11/16/2022]
Abstract
High-accurate and real-time localization is the fundamental and challenging task for autonomous driving in a dynamic traffic environment. This paper presents a coordinated positioning strategy that is composed of semantic information and probabilistic data association, which improves the accuracy of SLAM in dynamic traffic settings. First, the improved semantic segmentation network, building on Fast-SCNN, uses the Res2net module instead of the Bottleneck in the global feature extraction to further explore the multi-scale granular features. It achieves the balance between segmentation accuracy and inference speed, leading to consistent performance gains on the coordinated localization task of this paper. Second, a novel scene descriptor combining geometric, semantic, and distributional information is proposed. These descriptors are made up of significant features and their surroundings, which may be unique to a traffic scene, and are used to improve data association quality. Finally, a probabilistic data association is created to find the best estimate using a maximum measurement expectation model. This approach assigns semantic labels to landmarks observed in the environment and is used to correct false negatives in data association. We have evaluated our system with ORB-SLAM2 and DynaSLAM, the most advanced algorithms, to demonstrate its advantages. On the KITTI dataset, the results reveal that our approach outperforms other methods in dynamic traffic situations, especially in highly dynamic scenes, with sub-meter average accuracy.
Collapse
Affiliation(s)
- Kaixin Yang
- School of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China; (K.Y.); (W.Z.); (X.W.)
| | - Weiwei Zhang
- School of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China; (K.Y.); (W.Z.); (X.W.)
- School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China
- Shanghai Smart Vehicle Integration Innovation Center Co., Ltd., Shanghai 201620, China
| | - Chuanchang Li
- School of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China; (K.Y.); (W.Z.); (X.W.)
- Correspondence:
| | - Xiaolan Wang
- School of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China; (K.Y.); (W.Z.); (X.W.)
| |
Collapse
|
8
|
PFD-SLAM: A New RGB-D SLAM for Dynamic Indoor Environments Based on Non-Prior Semantic Segmentation. REMOTE SENSING 2022. [DOI: 10.3390/rs14102445] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Now, most existing dynamic RGB-D SLAM methods are based on deep learning or mathematical models. Abundant training sample data is necessary for deep learning, and the selection diversity of semantic samples and camera motion modes are closely related to the robust detection of moving targets. Furthermore, the mathematical models are implemented at the feature-level of segmentation, which is likely to cause sub or over-segmentation of dynamic features. To address this problem, different from most feature-level dynamic segmentation based on mathematical models, a non-prior semantic dynamic segmentation based on a particle filter is proposed in this paper, which aims to attain the motion object segmentation. Firstly, GMS and optical flow are used to calculate an inter-frame difference image, which is considered an observation measurement of posterior estimation. Then, a motion equation of a particle filter is established using Gaussian distribution. Finally, our proposed segmentation method is integrated into the front end of visual SLAM and establishes a new dynamic SLAM, PFD-SLAM. Extensive experiments on the public TUM datasets and real dynamic scenes are conducted to verify location accuracy and practical performances of PFD-SLAM. Furthermore, we also compare experimental results with several state-of-the-art dynamic SLAM methods in terms of two evaluation indexes, RPE and ATE. Still, we provide visual comparisons between the camera estimation trajectories and ground truth. The comprehensive verification and testing experiments demonstrate that our PFD-SLAM can achieve better dynamic segmentation results and robust performances.
Collapse
|
9
|
DGS-SLAM: A Fast and Robust RGBD SLAM in Dynamic Environments Combined by Geometric and Semantic Information. REMOTE SENSING 2022. [DOI: 10.3390/rs14030795] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
Abstract
Visual Simultaneous Localization and Mapping (VSLAM) is a prerequisite for robots to accomplish fully autonomous movement and exploration in unknown environments. At present, many impressive VSLAM systems have emerged, but most of them rely on the static world assumption, which limits their application in real dynamic scenarios. To improve the robustness and efficiency of the system in dynamic environments, this paper proposes a dynamic RGBD SLAM based on a combination of geometric and semantic information (DGS-SLAM). First, a dynamic object detection module based on the multinomial residual model is proposed, which executes the motion segmentation of the scene by combining the motion residual information of adjacent frames and the potential motion information of the semantic segmentation module. Second, a camera pose tracking strategy using feature point classification results is designed to achieve robust system tracking. Finally, according to the results of dynamic segmentation and camera tracking, a semantic segmentation module based on a semantic frame selection strategy is designed for extracting potential moving targets in the scene. Extensive evaluation in public TUM and Bonn datasets demonstrates that DGS-SLAM has higher robustness and speed than state-of-the-art dynamic RGB-D SLAM systems in dynamic scenes.
Collapse
|
10
|
Fu Q, Yu H, Wang X, Yang Z, He Y, Zhang H, Mian A. Fast ORB-SLAM Without Keypoint Descriptors. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:1433-1446. [PMID: 34951846 DOI: 10.1109/tip.2021.3136710] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Indirect methods for visual SLAM are gaining popularity due to their robustness to environmental variations. ORB-SLAM2 (Mur-Artal and Tardós, 2017) is a benchmark method in this domain, however, it consumes significant time for computing descriptors that never get reused unless a frame is selected as a keyframe. To overcome these problems, we present FastORB-SLAM which is light-weight and efficient as it tracks keypoints between adjacent frames without computing descriptors. To achieve this, a two stage descriptor-independent keypoint matching method is proposed based on sparse optical flow. In the first stage, we predict initial keypoint correspondences via a simple but effective motion model and then robustly establish the correspondences via pyramid-based sparse optical flow tracking. In the second stage, we leverage the constraints of the motion smoothness and epipolar geometry to refine the correspondences. In particular, our method computes descriptors only for keyframes. We test FastORB-SLAM on TUM and ICL-NUIM RGB-D datasets and compare its accuracy and efficiency to nine existing RGB-D SLAM methods. Qualitative and quantitative results show that our method achieves state-of-the-art accuracy and is about twice as fast as the ORB-SLAM2.
Collapse
|
11
|
VINS-dimc: A Visual-Inertial Navigation System for Dynamic Environment Integrating Multiple Constraints. ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION 2022. [DOI: 10.3390/ijgi11020095] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Most visual–inertial navigation systems (VINSs) suffer from moving objects and achieve poor positioning accuracy in dynamic environments. Therefore, to improve the positioning accuracy of VINS in dynamic environments, a monocular visual–inertial navigation system, VINS-dimc, is proposed. This system integrates various constraints on the elimination of dynamic feature points, which helps to improve the positioning accuracy of VINSs in dynamic environments. First, the motion model, computed from the inertial measurement unit (IMU) data, is subjected to epipolar constraint and flow vector bound (FVB) constraint to eliminate feature matching that deviates significantly from the motion model. This algorithm then combines multiple feature point matching constraints that avoid the lack of single constraints and make the system more robust and universal. Finally, VINS-dimc was proposed, which can adapt to a dynamic environment. Experiments show that the proposed algorithm could accurately eliminate the dynamic feature points on moving objects while preserving the static feature points. It is a great help for the positioning accuracy and robustness of VINSs, whether they are from self-collected data or public datasets.
Collapse
|
12
|
Sun L, Singh RP, Kanehiro F. Visual SLAM Framework Based on Segmentation with the Improvement of Loop Closure Detection in Dynamic Environments. JOURNAL OF ROBOTICS AND MECHATRONICS 2021. [DOI: 10.20965/jrm.2021.p1385] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Most simultaneous localization and mapping (SLAM) systems assume that SLAM is conducted in a static environment. When SLAM is used in dynamic environments, the accuracy of each part of the SLAM system is adversely affected. We term this problem as dynamic SLAM. In this study, we propose solutions for three main problems in dynamic SLAM: camera tracking, three-dimensional map reconstruction, and loop closure detection. We propose to employ geometry-based method, deep learning-based method, and the combination of them for object segmentation. Using the information from segmentation to generate the mask, we filter the keypoints that lead to errors in visual odometry and features extracted by the CNN from dynamic areas to improve the performance of loop closure detection. Then, we validate our proposed loop closure detection method using the precision-recall curve and also confirm the framework’s performance using multiple datasets. The absolute trajectory error and relative pose error are used as metrics to evaluate the accuracy of the proposed SLAM framework in comparison with state-of-the-art methods. The findings of this study can potentially improve the robustness of SLAM technology in situations where mobile robots work together with humans, while the object-based point cloud byproduct has potential for other robotics tasks.
Collapse
|
13
|
A Semantic SLAM System for Catadioptric Panoramic Cameras in Dynamic Environments. SENSORS 2021; 21:s21175889. [PMID: 34502780 PMCID: PMC8433785 DOI: 10.3390/s21175889] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/26/2021] [Revised: 08/16/2021] [Accepted: 08/25/2021] [Indexed: 11/30/2022]
Abstract
When a traditional visual SLAM system works in a dynamic environment, it will be disturbed by dynamic objects and perform poorly. In order to overcome the interference of dynamic objects, we propose a semantic SLAM system for catadioptric panoramic cameras in dynamic environments. A real-time instance segmentation network is used to detect potential moving targets in the panoramic image. In order to find the real dynamic targets, potential moving targets are verified according to the sphere’s epipolar constraints. Then, when extracting feature points, the dynamic objects in the panoramic image are masked. Only static feature points are used to estimate the pose of the panoramic camera, so as to improve the accuracy of pose estimation. In order to verify the performance of our system, experiments were conducted on public data sets. The experiments showed that in a highly dynamic environment, the accuracy of our system is significantly better than traditional algorithms. By calculating the RMSE of the absolute trajectory error, we found that our system performed up to 96.3% better than traditional SLAM. Our catadioptric panoramic camera semantic SLAM system has higher accuracy and robustness in complex dynamic environments.
Collapse
|
14
|
Dai W, Zhang Y, Zheng Y, Sun D, Li P. RGB‐D SLAM with moving object tracking in dynamic environments. IET CYBER-SYSTEMS AND ROBOTICS 2021. [DOI: 10.1049/csy2.12019] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022] Open
Affiliation(s)
- Weichen Dai
- State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University Hangzhou China
| | - Yu Zhang
- State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University Hangzhou China
| | - Yuxin Zheng
- State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University Hangzhou China
| | - Donglei Sun
- Centre for English Language Education, University of Nottingham Ningbo China Ningbo China
| | - Ping Li
- State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Zhejiang University Hangzhou China
| |
Collapse
|
15
|
Kim D, Pathak S, Moro A, Yamashita A, Asama H. Self-supervised optical flow derotation network for rotation estimation of a spherical camera. Adv Robot 2020. [DOI: 10.1080/01691864.2020.1857305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Affiliation(s)
- Dabae Kim
- Department of Precision Engineering, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| | - Sarthak Pathak
- Department of Precision Engineering, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| | - Alessandro Moro
- Department of Precision Engineering, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| | - Atsushi Yamashita
- Department of Precision Engineering, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| | - Hajime Asama
- Department of Precision Engineering, Graduate School of Engineering, The University of Tokyo, Tokyo, Japan
| |
Collapse
|
16
|
Kang Z, Zou W. Improving accuracy of VI-SLAM with fish-eye camera based on biases of map points. Adv Robot 2020. [DOI: 10.1080/01691864.2020.1815573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Affiliation(s)
- Zhaobing Kang
- Institute of Automation Chinese Academy of Sciences, University of Chinese Academy of Sciences, Beijing, People's Republic of China
| | - Wei Zou
- Institute of Automation, Chinese Academy of Sciences, Beijing, People's Republic of China
- TianJin Intelligent Tech. Institute of CASIA Co. Ltd, Tianjin, People's Republic of China
| |
Collapse
|
17
|
Sun Y, Zuo W, Liu M. See the Future: A Semantic Segmentation Network Predicting Ego-Vehicle Trajectory With a Single Monocular Camera. IEEE Robot Autom Lett 2020. [DOI: 10.1109/lra.2020.2975414] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
18
|
An Improved Deep Residual Network-Based Semantic Simultaneous Localization and Mapping Method for Monocular Vision Robot. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2020; 2020:7490840. [PMID: 32104171 PMCID: PMC7035522 DOI: 10.1155/2020/7490840] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/10/2019] [Revised: 12/31/2019] [Accepted: 01/14/2020] [Indexed: 11/18/2022]
Abstract
The robot simultaneous localization and mapping (SLAM) is a very important and useful technology in the robotic field. However, the environmental map constructed by the traditional visual SLAM method contains little semantic information, which cannot satisfy the needs of complex applications. The semantic map can deal with this problem efficiently, which has become a research hot spot. This paper proposed an improved deep residual network- (ResNet-) based semantic SLAM method for monocular vision robots. In the proposed approach, an improved image matching algorithm based on feature points is presented, to enhance the anti-interference ability of the algorithm. Then, the robust feature point extraction method is adopted in the front-end module of the SLAM system, which can effectively reduce the probability of camera tracking loss. In addition, the improved key frame insertion method is introduced in the visual SLAM system to enhance the stability of the system during the turning and moving of the robot. Furthermore, an improved ResNet model is proposed to extract the semantic information of the environment to complete the construction of the semantic map of the environment. Finally, various experiments are conducted and the results show that the proposed method is effective.
Collapse
|