RGBTSDF: An Efficient and Simple Method for Color Truncated Signed Distance Field (TSDF) Volume Fusion Based on RGB-D Images
-
Published:2024-08-29
Issue:17
Volume:16
Page:3188
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Li Yunqiang12ORCID, Huang Shuowen1ORCID, Chen Ying2, Ding Yong2, Zhao Pengcheng1ORCID, Hu Qingwu1ORCID, Zhang Xujie13
Affiliation:
1. School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China 2. Wuhan Ruler Technology, Wuhan 430079, China 3. Hubei Luojia Laboratory, Wuhan 430079, China
Abstract
RGB-D image mapping is an important tool in applications such as robotics, 3D reconstruction, autonomous navigation, and augmented reality (AR). Efficient and reliable mapping methods can improve the accuracy, real-time performance, and flexibility of sensors in various fields. However, the currently widely used Truncated Signed Distance Field (TSDF) still suffers from the problem of inefficient memory management, making it difficult to directly use it for large-scale 3D reconstruction. In order to address this problem, this paper proposes a highly efficient and accurate TSDF voxel fusion method, RGBTSDF. First, based on the sparse characteristics of the volume, an improved grid octree is used to manage the whole scene, and a hard coding method is proposed for indexing. Second, during the depth map fusion process, the depth map is interpolated to achieve a more accurate voxel fusion effect. Finally, a mesh extraction method with texture constraints is proposed to overcome the effects of noise and holes and improve the smoothness and refinement of the extracted surface. We comprehensively evaluate RGBTSDF and similar methods through experiments on public datasets and the datasets collected by commercial scanning devices. Experimental results show that RGBTSDF requires less memory and can achieve real-time performance experience using only the CPU. It also improves fusion accuracy and achieves finer grid details.
Funder
State Key Laboratory of Geo-Information Engineering Wuhan Ruler Technology Foundation National Natural Science Foundation of China
Reference36 articles.
1. Izadi, S., Kim, D., Hilliges, O., Molyneaux, D., Newcombe, R., Kohli, P., Shotton, J., Hodges, S., Freeman, D., and Davison, A. (2011, January 16–19). Kinectfusion: Real-Time 3D Reconstruction and Interaction Using a Moving Depth Camera. Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology, Santa Barbara, CA, USA. 2. Curless, B., and Levoy, M. (1996, January 4–9). A Volumetric Method for Building Complex Models from Range Images. Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA. 3. Cao, Y., Liu, Z., Kuang, Z., Kobbelt, L., and Hu, S. (2018, January 8–14). Learning to Reconstruct High-Quality 3D Shapes with Cascaded Fully Convolutional Networks. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany. 4. Isler, S., Sabzevari, R., Delmerico, J., and Scaramuzza, D. (2016, January 16–21). An Information Gain Formulation for Active Volumetric 3D Reconstruction. Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden. 5. Hinzmann, T., Schönberger, J.L., Pollefeys, M., and In Siegwart, R. (2017, January 12–15). Mapping on the Fly: Real-Time 3D Dense Reconstruction, Digital Surface Map and Incremental Orthomosaic Generation for Unmanned Aerial Vehicles. Proceedings of the Field and Service Robotics: Results of the 11th International Conference, Zurich, Switzerland.
|
|