Affiliation:
1. College of Mechanical Engineering and Automation, Huaqiao University, Xiamen 361021, China
2. Fujian Key Laboratory of Green Intelligent Drive and Transmission for Mobile Machinery, Xiamen 361021, China
Abstract
Simultaneous localization and mapping (SLAM), as a key task of unmanned vehicles for construction machinery, is of great significance for later path planning and control. Construction tasks in the engineering field are mostly carried out in bridges, tunnels, open fields, etc. The prominent features of these environments are high scene similarity, few geometric features, and large-scale repetitive texture information, which is prone to sensor detection degradation. This leads to positioning drift and map building failure. The traditional method of motion estimation and 3D reconstruction uses a single sensor, which lacks enough information, has poor adaptability to the environment, and cannot guarantee good positioning accuracy and robustness in complex environments. Currently, the strategy of multi-sensor fusion is proven to be an effective solution and is widely studied. This paper proposes a SLAM framework that integrates LiDAR, IMU, and camera. It tightly couples the texture information observed by camera, the geometric information scanned by LiDAR, and the measured value of IMU, allowing visual-inertial odometry (VIO) and LiDAR-inertial odometry (LIO) common implementation. The LIO subsystem extracts point cloud features and matches them with the global map. The obtained pose estimation can be used for the initialization of the VIO subsystem. The VIO system uses direct method to minimize the photometric error and IMU measurement error between images to estimate the pose of the robot and the geometric structure of the scene. The two subsystems assist each other to perform pose estimation, and can operate normally even when any subsystem fails. A factor graph is used to combine all constraints to achieve global pose optimization. Keyframe and sliding window strategies are used to ensure real-time performance. Through real-vehicle testing, the system can perform incremental and real-time state estimation and reconstruct a dense 3D point cloud map, which can effectively solve the problems of positioning drift and mapping failure in the lack of geometric features or challenging construction environments.
Funder
Fujian University industry university research joint innovation project plan
Industry Cooperation of Major Science and Technology Project of Fujian Province
Xiamen Major Science and Technology Plan Projects
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Reference38 articles.
1. Graeter, J., Wilczynski, A., and Lauer, M. (2018, January 1–5). Limo: Lidar-monocular visual odometry. Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain.
2. Dvl-slam: Sparse depth enhanced direct visual-lidar slam;Shin;Auton. Robot.,2020
3. Shin, Y., Park, Y.S., and Kim, A. (2018, January 21–25). Direct visual slam using sparse depth for camera-lidar system. Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia.
4. Silva, V.D., Roche, J., and Kondoz, A. (2017). Fusion of lidar and camera sensor data for environment sensing in driverless vehicles. arXiv.
5. River mapping from a flying robot: State estimation, river detection, and obstacle mapping;Scherer;Auton. Robot.,2012