A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition
-
Published:2024-08-23
Issue:17
Volume:16
Page:3121
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Liu Hehao1ORCID, Li Dong1, Zhang Ming2, Wan Jun1, Liu Shuang1ORCID, Zhu Hanying1, Liu Qinghua3
Affiliation:
1. School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China 2. South-West Institute of Electronics and Telecommunication Technology, Chengdu 610041, China 3. Guangxi Key Laboratory of Wireless Wideband Communication and Signal Processing, Guilin 541004, China
Abstract
With the continuous progress in drone and materials technology, numerous bionic drones have been developed and employed in various fields. These bionic drones are designed to mimic the shape of birds, seamlessly blending into the natural environment and reducing the likelihood of detection. However, such a high degree of similarity also poses significant challenges in accurately distinguishing between real birds and bionic drones. Existing methods attempt to recognize both using optical images, but the visual similarity often results in poor recognition accuracy. To alleviate this problem, in this paper, we propose a cross-modal semantic alignment and feature fusion (CSAFF) network to improve the recognition accuracy of bionic drones. CSAFF aims to introduce motion behavior information as an auxiliary cue to improve discriminability. Specifically, a semantic alignment module (SAM) was designed to explore the consistent semantic information between cross-modal data and provide more semantic cues for the recognition of bionic drones and birds. Then, a feature fusion module (FFM) was developed to fully integrate cross-modal information, which effectively enhances the representability of these features. Extensive experiments were performed on datasets containing bionic drones and birds, and the experimental results consistently show the effectiveness of the proposed CSAFF method in identifying bionic drones and birds.
Funder
National Natural Science Foundation of China the Defense Industrial Technology Development Program the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory Sichuan Science and Technology Program the Engineering Research Center of Mobile Communications, Ministry of Education the Natural Science Foundation of Chongqing, China
Reference57 articles.
1. Top 11 technologies of the decade;Ross;IEEE Spectr.,2010 2. Avola, D., Cannistraci, I., Cascio, M., Cinque, L., Diko, A., Fagioli, A., Foresti, G.L., Lanzino, R., Mancini, M., and Mecca, A. (2022). A novel GAN-based anomaly detection and localization method for aerial video surveillance at low altitude. Remote Sens., 14. 3. Ritchie, M., Fioranelli, F., Griffiths, H., and Torvik, B. (2015, January 27–30). Micro-drone RCS analysis. Proceedings of the IEEE Radar Conference, Johannesburg, South Africa. 4. In-flight RCS measurements of drones and birds at K-band and W-band;Rahman;IET Radar Sonar Navig.,2019 5. Rojhani, N., and Shaker, G. (2024). Comprehensive Review: Effectiveness of MIMO and Beamforming Technologies in Detecting Low RCS UAVs. Remote Sens., 16.
|
|