An Asymmetric Feature Enhancement Network for Multiple Object Tracking of Unmanned Aerial Vehicle
-
Published:2023-12-23
Issue:1
Volume:16
Page:70
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Ma Jianbo1234, Liu Dongxu123ORCID, Qin Senlin1234, Jia Ge123, Zhang Jianlin123ORCID, Xu Zhiyong123
Affiliation:
1. National Key Laboratory of Optical Field Manipulation Science and Technology, Chinese Academy of Sciences, Chengdu 610209, China 2. Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China 3. Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China 4. University of Chinese Academy of Sciences, Beijing 100049, China
Abstract
Multiple object tracking (MOT) in videos captured by unmanned aerial vehicle (UAV) is a fundamental aspect of computer vision. Recently, the one-shot tracking paradigm integrates the detection and re-identification (ReID) tasks, striking a balance between tracking accuracy and inference speed. This paradigm alleviates task conflicts and achieves remarkable results through various feature decoupling methods. However, in challenging scenarios like drone movements, lighting changes and object occlusion, it still encounters issues with detection failures and identity switches. In addition, traditional feature decoupling methods directly employ channel-based attention to decompose the detection and ReID branches, without a meticulous consideration of the specific requirements of each branch. To address the above problems, we introduce an asymmetric feature enhancement network with a global coordinate-aware enhancement (GCAE) module and an embedding feature aggregation (EFA) module, aiming to optimize the two branches independently. On the one hand, we develop the GCAE module for the detection branch, which effectively merges rich semantic information within the feature space to improve detection accuracy. On the other hand, we introduce the EFA module for the ReID branch, which highlights the significance of pixel-level features and acquires discriminative identity embedding through a local feature aggregation strategy. By efficiently incorporating the GCAE and EFA modules into the one-shot tracking pipeline, we present a novel MOT framework, named AsyUAV. Extensive experiments have demonstrated the effectiveness of our proposed AsyUAV. In particular, it achieves a MOTA of 38.3% and IDF1 of 51.7% on VisDrone2019, and a MOTA of 48.0% and IDF1 of 67.5% on UAVDT, outperforming existing state-of-the-art trackers.
Funder
National Natural Science Foundation of China
Subject
General Earth and Planetary Sciences
Reference55 articles.
1. Wu, Z., Liu, Q., Zhou, S., Qiu, S., Zhang, Z., and Zeng, Y. (2023). Learning Template-Constraint Real-Time Siamese Tracker for Drone AI Devices via Concatenation. Drones, 7. 2. Avola, D., Cinque, L., Diko, A., Fagioli, A., Foresti, G.L., Mecca, A., Pannone, D., and Piciarelli, C. (2021). MS-Faster R-CNN: Multi-Stream Backbone for Improved Faster R-CNN Object Detection and Aerial Tracking from UAV Images. Remote Sens., 13. 3. Li, X., and Wu, J. (2022). Extracting High-Precision Vehicle Motion Data from Unmanned Aerial Vehicle Video Captured under Various Weather Conditions. Remote Sens., 14. 4. Wang, G., Song, M., and Hwang, J.N. (2022). Recent advances in embedding methods for multi-object tracking: A survey. arXiv. 5. Varga, L.A., Koch, S., and Zell, A. (2022). Comprehensive Analysis of the Object Detection Pipeline on UAVs. Remote Sens., 14.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|