Author:
Yin Wenji,Peng Yueping,Ye Zecong,Liu Wenchao
Abstract
Vehicle re-identification research under surveillance cameras has yielded impressive results. However, the challenge of Unmanned Aerial Vehicle (UAV)-based vehicle re-identification (ReID) presents a high degree of flexibility, mainly due to complicated shooting angles, occlusions, low discrimination of top-down features, and significant changes in vehicle scales. To address this, we propose a novel Dual Mixing Attention Network (DMANet) to extract discriminative features robust to variations in viewpoint. Specifically, we first present a plug-and-play Dual Mixing Attention Module (DMAM) to capture pixel-level pairwise relationships and channel dependencies, where DMAM is composed of Spatial Mixing Attention (SMA) and Channel Mixing Attention (CMA). First, the original feature is divided according to the dimensions of spatial and channel to obtain multiple subspaces. Then, a learnable weight is applied to capture the dependencies between local features in the mixture space. Finally, the features extracted from all subspaces are aggregated to promote their comprehensive feature interaction. Moreover, the DMAM can be readily inserted into backbone networks at any depth to improve vehicle discrimination. The experiments show that the proposed structure gains a better performance compared with the representative methods in the UAV-based vehicle ReID task. Our code and models will be publicly released.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献