Visual Object Tracking Based on the Motion Prediction and Block Search in UAV Videos
Author:
Sun Lifan123, Li Xinxiang1ORCID, Yang Zhe4, Gao Dan1
Affiliation:
1. School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China 2. Longmen Laboratory, Luoyang 471000, China 3. Henan Academy of Sciences, Zhengzhou 450046, China 4. Xiaomi Technology Co., Ltd., Beijing 100102, China
Abstract
With the development of computer vision and Unmanned Aerial Vehicles (UAVs) technology, visual object tracking has become an indispensable core technology for UAVs, and it has been widely used in both civil and military fields. Visual object tracking from the UAV perspective experiences interference from various complex conditions such as background clutter, occlusion, and being out of view, which can easily lead to tracking drift. Once tracking drift occurs, it will lead to almost complete failure of the subsequent tracking. Currently, few trackers have been designed to solve the tracking drift problem. Thus, this paper proposes a tracking algorithm based on motion prediction and block search to address the tracking drift problem caused by various complex conditions. Specifically, when the tracker experiences tracking drift, we first use a Kalman filter to predict the motion state of the target, and then use a block search module to relocate the target. In addition, to improve the tracker’s ability to adapt to changes in the target’s appearance and the environment, we propose a dynamic template updating network (DTUN) that allows the tracker to make appropriate template decisions based on various tracking conditions. We also introduce three tracking evaluation metrics: namely, average peak correlation energy, size change ratio, and tracking score. They serve as prior information for tracking status identification in the DTUN and the block prediction module. Extensive experiments and comparisons with many competitive algorithms on five aerial benchmarks, UAV20L, UAV123, UAVDT, DTB70, and VisDrone2018-SOT, demonstrate that our method achieves significant performance improvements. Especially in UAV20L long-term tracking, our method outperforms the baseline in terms of success rate and accuracy by 19.1% and 20.8%, respectively. This demonstrates the superior performance of our method in the task of long-term tracking from the UAV perspective, and we achieve a real-time speed of 43 FPS.
Funder
National Natural Science Foundation of China Aeronautical Science Foundation of China Natural Science Foundation of Henan Province, China
Reference50 articles.
1. Yeom, S. (2024). Thermal Image Tracking for Search and Rescue Missions with a Drone. Drones, 8. 2. Han, Y., Yu, X., Luan, H., and Suo, J. (2024). Event-Assisted Object Tracking on High-Speed Drones in Harsh Illumination Environment. Drones, 8. 3. Chen, Q., Liu, J., Liu, F., Xu, F., and Liu, C. (2024). Lightweight Spatial-Temporal Contextual Aggregation Siamese Network for Unmanned Aerial Vehicle Tracking. Drones, 8. 4. Memon, S.A., Son, H., Kim, W.G., Khan, A.M., Shahzad, M., and Khan, U. (2023). Tracking Multiple Unmanned Aerial Vehicles through Occlusion in Low-Altitude Airspace. Drones, 7. 5. Gao, Y., Gan, Z., Chen, M., Ma, H., and Mao, X. (2023). Hybrid Dual-Scale Neural Network Model for Tracking Complex Maneuvering UAVs. Drones, 8.
|
|