MATI: Multimodal Adaptive Tracking Integrator for Robust Visual Object Tracking
Author:
Li Kai12, Cai Lihua1, He Guangjian1, Gong Xun1
Affiliation:
1. Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China 2. University of Chinese Academy of Sciences, Beijing 100049, China
Abstract
Visual object tracking, pivotal for applications like earth observation and environmental monitoring, encounters challenges under adverse conditions such as low light and complex backgrounds. Traditional tracking technologies often falter, especially when tracking dynamic objects like aircraft amidst rapid movements and environmental disturbances. This study introduces an innovative adaptive multimodal image object-tracking model that harnesses the capabilities of multispectral image sensors, combining infrared and visible light imagery to significantly enhance tracking accuracy and robustness. By employing the advanced vision transformer architecture and integrating token spatial filtering (TSF) and crossmodal compensation (CMC), our model dynamically adjusts to diverse tracking scenarios. Comprehensive experiments conducted on a private dataset and various public datasets demonstrate the model’s superior performance under extreme conditions, affirming its adaptability to rapid environmental changes and sensor limitations. This research not only advances visual tracking technology but also offers extensive insights into multisource image fusion and adaptive tracking strategies, establishing a robust foundation for future enhancements in sensor-based tracking systems.
Funder
National Natural Science Foundation of China
Reference44 articles.
1. Wang, D., Wang, J.G., and Xu, K. (2021). Deep learning for object detection, classification and tracking in industry applications. Sensors, 21. 2. Danelljan, M., Hager, G., Shahbaz Khan, F., and Felsberg, M. (2015, January 7–13). Learning Spatially Regularized Correlation Filters for Visual Tracking. Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile. 3. Kiani Galoogahi, H., Fagg, A., and Lucey, S. (2017, January 22–29). Learning Background-Aware Correlation Filters for Visual Tracking. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy. 4. Bertinetto, L., Valmadre, J., Henriques, J.F., Vedaldi, A., and Torr, P.H. (15–16, January 8–10). Fully-Convolutional Siamese Networks for Object Tracking. Proceedings of the Computer Vision–ECCV 2016 Workshops, Amsterdam, The Netherlands. Proceedings, Part II 14. 5. Li, B., Yan, J., Wu, W., Zhu, Z., and Hu, X. (2018, January 18–22). High Performance Visual Tracking with Siamese Region Proposal Network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.
|
|