Voxel- and Bird’s-Eye-View-Based Semantic Scene Completion for LiDAR Point Clouds
-
Published:2024-06-21
Issue:13
Volume:16
Page:2266
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Liang Li1ORCID, Akhtar Naveed2ORCID, Vice Jordan1ORCID, Mian Ajmal1ORCID
Affiliation:
1. Department of Computer Science and Software Engineering, The University of Western Australia, 35 Stirling Hwy, Crawley, WA 6009, Australia 2. School of Computing and Information Systems, The University of Melbourne, Parkville, VIC 3052, Australia
Abstract
Semantic scene completion is a crucial outdoor scene understanding task that has direct implications for technologies like autonomous driving and robotics. It compensates for unavoidable occlusions and partial measurements in LiDAR scans, which may otherwise cause catastrophic failures. Due to the inherent complexity of this task, existing methods generally rely on complex and computationally demanding scene completion models, which limits their practicality in downstream applications. Addressing this, we propose a novel integrated network that combines the strengths of 3D and 2D semantic scene completion techniques for efficient LiDAR point cloud scene completion. Our network leverages a newly devised lightweight multi-scale convolutional block (MSB) to efficiently aggregate multi-scale features, thereby improving the identification of small and distant objects. It further utilizes a layout-aware semantic block (LSB), developed to grasp the overall layout of the scene to precisely guide the reconstruction and recognition of features. Moreover, we also develop a feature fusion module (FFM) for effective interaction between the data derived from two disparate streams in our network, ensuring a robust and cohesive scene completion process. Extensive experiments with the popular SemanticKITTI dataset demonstrate that our method achieves highly competitive performance, with an mIoU of 35.7 and an IoU of 51.4. Notably, the proposed method achieves an mIoU improvement of 2.6 % compared to previous methods.
Funder
Australian Research Council Future Fellowship Award
Reference56 articles.
1. Song, S., Yu, F., Zeng, A., Chang, A.X., Savva, M., and Funkhouser, T. (2017, January 21–26). Semantic scene completion from a single depth image. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA. 2. Roldao, L., de Charette, R., and Verroust-Blondet, A. (2020, January 25–28). Lmscnet: Lightweight multiscale 3d semantic completion. Proceedings of the 2020 International Conference on 3D Vision (3DV), Fukuoka, Japan. 3. Yan, X., Gao, J., Li, J., Zhang, R., Li, Z., Huang, R., and Cui, S. (2021, January 2–9). Sparse single sweep lidar point cloud segmentation via learning contextual shape priors from scene completion. Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada. 4. Cheng, R., Agia, C., Ren, Y., Li, X., and Bingbing, L. (2021, January 8–11). S3cnet: A sparse semantic scene completion network for lidar point clouds. Proceedings of the Conference on Robot Learning, PMLR, London, UK. 5. Yang, X., Zou, H., Kong, X., Huang, T., Liu, Y., Li, W., Wen, F., and Zhang, H. (October, January 27). Semantic segmentation-assisted scene completion for lidar point clouds. Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic.
|
|