MInet: A Novel Network Model for Point Cloud Processing by Integrating Multi-Modal Information
Author:
Wang Yuhao1ORCID, Zuo Yong1, Du Zhihua1, Song Xiaohan1, Luo Tian1, Hong Xiaobin1, Wu Jian1
Affiliation:
1. School of Electronic Engineering, Beijing University of Post and Telecommunications, Beijing 100876, China
Abstract
Three-dimensional LiDAR systems that capture point cloud data enable the simultaneous acquisition of spatial geometry and multi-wavelength intensity information, thereby paving the way for three-dimensional point cloud recognition and processing. However, due to the irregular distribution, low resolution of point clouds, and limited spatial recognition accuracy in complex environments, inherent errors occur in classifying and segmenting the acquired target information. Conversely, two-dimensional visible light images provide real-color information, enabling the distinction of object contours and fine details, thus yielding clear, high-resolution images when desired. The integration of two-dimensional information with point clouds offers complementary advantages. In this paper, we present the incorporation of two-dimensional information to form a multi-modal representation. From this, we extract local features to establish three-dimensional geometric relationships and two-dimensional color relationships. We introduce a novel network model, termed MInet (Multi-Information net), which effectively captures features relating to both two-dimensional color and three-dimensional pose information. This enhanced network model improves feature saliency, thereby facilitating superior segmentation and recognition tasks. We evaluate our MInet architecture using the ShapeNet and ThreeDMatch datasets for point cloud segmentation, and the Stanford dataset for object recognition. The robust results, coupled with quantitative and qualitative experiments, demonstrate the superior performance of our proposed method in point cloud segmentation and object recognition tasks.
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference31 articles.
1. Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions;Liu;IEEE Trans. Intell. Transp. Syst.,2021 2. Jiang, J., Liu, D., Gu, J., and Süsstrunk, S. (2013, January 15–17). What is the space of spectral sensitivity functions for digital color cameras?. Proceedings of the 2013 IEEE Workshop on Applications of Computer Vision (WACV), Clearwater Beach, FL, USA. 3. Villa, F., Severini, F., Madonini, F., and Zappa, F. (2021). SPADs and SiPMs arrays for long-range high-speed light detection and ranging (LiDAR). Sensors, 21. 4. Cheng, L., Chen, S., Liu, X., Xu, H., Wu, Y., Li, M., and Chen, Y. (2018). Registration of laser scanning point clouds: A review. Sensors, 18. 5. Schumann, O., Hahn, M., Dickmann, J., and Wöhler, C. (2018, January 10–13). Semantic segmentation on radar point clouds. Proceedings of the 2018 21st International Conference on Information Fusion (FUSION), Cambridge, UK.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|