DIO-SLAM: A Dynamic RGB-D SLAM Method Combining Instance Segmentation and Optical Flow
Author:
He Lang1ORCID, Li Shiyun1, Qiu Junting2, Zhang Chenhaomin1
Affiliation:
1. Faculty of Mechanical and Electrical Engineering, Kunming University of Science and Technology, Kunming 650500, China 2. College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou 310014, China
Abstract
Feature points from moving objects can negatively impact the accuracy of Visual Simultaneous Localization and Mapping (VSLAM) algorithms, while detection or semantic segmentation-based VSLAM approaches often fail to accurately determine the true motion state of objects. To address this challenge, this paper introduces DIO-SLAM: Dynamic Instance Optical Flow SLAM, a VSLAM system specifically designed for dynamic environments. Initially, the detection thread employs YOLACT (You Only Look At CoefficienTs) to distinguish between rigid and non-rigid objects within the scene. Subsequently, the optical flow thread estimates optical flow and introduces a novel approach to capture the optical flow of moving objects by leveraging optical flow residuals. Following this, an optical flow consistency method is implemented to assess the dynamic nature of rigid object mask regions, classifying them as either moving or stationary rigid objects. To mitigate errors caused by missed detections or motion blur, a motion frame propagation method is employed. Lastly, a dense mapping thread is incorporated to filter out non-rigid objects using semantic information, track the point clouds of rigid objects, reconstruct the static background, and store the resulting map in an octree format. Experimental results demonstrate that the proposed method surpasses current mainstream dynamic VSLAM techniques in both localization accuracy and real-time performance.
Reference45 articles.
1. RLD-SLAM: A Robust Lightweight VI-SLAM for Dynamic Environments Leveraging Semantics and Motion Information;Zheng;IEEE Trans. Ind. Electron.,2024 2. Jia, G., Li, X., Zhang, D., Xu, W., Lv, H., Shi, Y., and Cai, M. (2022). Visual-SLAM Classical Framework and Key Techniques: A Review. Sensors, 22. 3. Chen, W., Shang, G., Ji, A., Zhou, C., Wang, X., Xu, C., Li, Z., and Hu, K. (2022). An Overview on Visual SLAM: From Tradition to Semantic. Remote Sens., 14. 4. Macario Barros, A., Michel, M., Moline, Y., Corre, G., and Carrel, F. (2022). A Comprehensive Survey of Visual SLAM Algorithms. Robotics, 11. 5. Tourani, A., Bavle, H., Sanchez-Lopez, J.L., and Voos, H. (2022). Visual SLAM: What Are the Current Trends and What to Expect?. Sensors, 22.
|
|