Boost Correlation Features with 3D-MiIoU-Based Camera-LiDAR Fusion for MODT in Autonomous Driving
-
Published:2023-02-04
Issue:4
Volume:15
Page:874
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Zhang Kunpeng1ORCID, Liu Yanheng12, Mei Fang12ORCID, Jin Jingyi1, Wang Yiming1
Affiliation:
1. College of Computer Science and Technology, Jilin University, Changchun 130012, China 2. Key Laboratory of Symbol Computation and Knowledge Engineering of the Ministry of Education, Jilin University, Changchun 130012, China
Abstract
Three-dimensional (3D) object tracking is critical in 3D computer vision. It has applications in autonomous driving, robotics, and human–computer interaction. However, methods for using multimodal information among objects to increase multi-object detection and tracking (MOT) accuracy remain a critical focus of research. Therefore, we present a multimodal MOT framework for autonomous driving boost correlation multi-object detection and tracking (BcMODT) in this research study to provide more trustworthy features and correlation scores for real-time detection tracking using both camera and LiDAR measurement data. Specifically, we propose an end-to-end deep neural network using 2D and 3D data for joint object detection and association. A new 3D mixed IoU (3D-MiIoU) computational module is also developed to acquire more precise geometric affinity by increasing the aspect ratio and length-to-height ratio between linked frames. Meanwhile, a boost correlation feature (BcF) module is proposed for the affinity calculation of the appearance of similar objects, which comprises an appearance affinity calculation module for similar objects in adjacent frames that are calculated directly using the feature distance and feature direction’s similarity. The KITTI tracking benchmark shows that our method outperforms other methods with respect to tracking accuracy.
Funder
Science and Technology Development Plan Project of Jilin Province National Natural Science Foundation of China
Subject
General Earth and Planetary Sciences
Reference40 articles.
1. Weng, X., Wang, Y., Man, Y., and Kitani, K.M. (2020, January 13–19). Gnn3dmot: Graph neural network for 3d multi-object tracking with 2d-3d multi-feature learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 2. Wu, J., Cao, J., Song, L., Wang, Y., Yang, M., and Yuan, J. (2021, January 20–25). Track to detect and segment: An online multi-object tracker. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA. 3. Leibe, B., Schindler, K., and Van Gool, L. (2007, January 14–21). Coupled detection and trajectory estimation for multi-object tracking. Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil. 4. Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets, methods, and challenges;Feng;IEEE Trans. Intell. Transp. Syst.,2020 5. Kim, A., Ošep, A., and Leal-Taixé, L. (June, January 30). Eagermot: 3d multi-object tracking via sensor fusion. Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China.
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|