Stage-Aware Interaction Network for Point Cloud Completion
-
Published:2024-08-20
Issue:16
Volume:13
Page:3296
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Affiliation:
1. School of Mechanical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
Abstract
Point cloud completion aims to restore full shapes of objects from partial scans, and a typical network pipeline is AutoEncoder, which has coarse-to-fine refinement modules. Although existing approaches using this kind of architecture achieve promising results, they usually neglect the usage of shallow geometry features in partial inputs and the fusion of multi-stage features in the upsampling process, which prevents network performances from further improving. Therefore, in this paper, we propose a new method with dense interactions between different encoding and decoding steps. First, we introduce the Decoupled Multi-head Transformer (DMT), which implements and integrates semantic prediction and resolution upsampling in a unified network module, which serves as a primary ingredient in our pipeline. Second, we propose an Encoding-aware Coarse Decoder (ECD) that compactly makes the top–down shape-decoding process interact with the bottom–up feature-encoding process to utilize both shallow and deep features of partial inputs for coarse point cloud generation. Third, we design a Stage-aware Refinement Group (SRG), which comprehensively understands local semantics from densely connected features across different decoding stages and gradually upsamples point clouds based on them. In general, the key contributions of our method are the DMT for joint semantic-resolution generation, the ECD for multi-scale feature fusion-based shape decoding, and the SRG for stage-aware shape refinement. Evaluations on two synthetic and three real-world datasets illustrate that our method achieves competitive performances compared with existing approaches.
Funder
National Natural Science Foundation of China
Reference64 articles.
1. Geiger, A., Lenz, P., and Urtasun, R. (2012, January 16–21). Are we ready for autonomous driving? The KITTI vision benchmark suite. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA. 2. Dai, A., Chang, A.X., Savva, M., Halber, M., Funkhouser, T., and Nießner, M. (2017, January 21–26). ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA. 3. Yuan, W., Khot, T., Held, D., Mertz, C., and Hebert, M. (2018, January 5–8). PCN: Point Completion Network. Proceedings of the International Conference on 3D Vision (3DV), Verona, Italy. 4. Wang, X., Ang, M.H., and Lee, G.H. (2020, January 13–19). Cascaded Refinement Network for Point Cloud Completion. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA. 5. Xie, H., Yao, H., Zhou, S., Mao, J., Zhang, S., and Sun, W. (2020, January 23–28). GRNet: Gridding Residual Network for Dense Point Cloud Completion. Proceedings of the European Conference on Computer Vision (ECCV), Online.
|
|