A Registration Method of Overlap Aware Point Clouds Based on Transformer-to-Transformer Regression
-
Published:2024-05-25
Issue:11
Volume:16
Page:1898
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Zhao Yafei1ORCID, Chen Lineng2ORCID, Zhou Quanchen1ORCID, Zuo Jiabao1ORCID, Wang Huan1ORCID, Ren Mingwu1ORCID
Affiliation:
1. School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China 2. School of Electronic and Information Engineering, Guangxi Normal University, Guilin 541004, China
Abstract
Transformer has recently become widely adopted in point cloud registration. Nevertheless, Transformer is unsuitable for handling dense point clouds due to resource constraints and the sheer volume of data. We propose a method for directly regressing the rigid relative transformation of dense point cloud pairs. Specifically, we divide the dense point clouds into blocks according to the down-sampled superpoints. During training, we randomly select point cloud blocks with varying overlap ratios, and during testing, we introduce the overlap-aware Rotation-Invariant Geometric Transformer Cross-Encoder (RIG-Transformer), which predicts superpoints situated within the common area of the point cloud pairs. The dense points corresponding to the superpoints are inputted into the Transformer Cross-Encoder to estimate their correspondences. Through the fusion of our RIG-Transformer and Transformer Cross-Encoder, we propose Transformer-to-Transformer Regression (TTReg), which leverages dense point clouds from overlapping regions for both training and testing phases, calculating the relative transformation of the dense points by using the predicted correspondences without random sample consensus (RANSAC). We have evaluated our method on challenging benchmark datasets, including 3DMatch, 3DLoMatch, ModelNet, and ModelLoNet, demonstrating up to a 7.2% improvement in registration recall. The improvements are attributed to our RIG-Transformer module and regression mechanism, which makes the features of superpoints more discriminative.
Funder
National Natural Science Foundation of China Natural Science Foundation of the Higher Education Institutions of Jiangsu Province Qing Lan Project of Jiangsu Province Cultivation Object of Major Scientific Research Project of CZIMT Nanjing University of Science and Technology
Reference34 articles.
1. Chen, Y., Mei, Y., Yu, B., Xu, W., Wu, Y., Zhang, D., and Yan, X. (2023). A robust multi-local to global with outlier filtering for point cloud registration. Remote Sens., 15. 2. Sumetheeprasit, B., Rosales Martinez, R., Paul, H., and Shimonomura, K. (2024). Long-range 3D reconstruction based on flexible configuration stereo vision using multiple aerial robots. Remote Sens., 16. 3. Choy, C., Park, J., and Koltun, V. (November, January 27). Fully convolutional geometric features. Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea. 4. Han, T., Zhang, R., Kan, J., Dong, R., Zhao, X., and Yao, S. (2024). A point cloud registration framework with color information integration. Remote Sens., 16. 5. Mei, G., Tang, H., Huang, X., Wang, W., Liu, J., Zhang, J., Van Gool, L., and Wu, Q. (2023, January 17–24). Unsupervised deep probabilistic approach for partial point cloud registration. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.
|
|