CCDN-DETR: A Detection Transformer Based on Constrained Contrast Denoising for Multi-Class Synthetic Aperture Radar Object Detection
Author:
Zhang Lei1, Zheng Jiachun1, Li Chaopeng1, Xu Zhiping1ORCID, Yang Jiawen1, Wei Qiuxin2, Wu Xinyi1
Affiliation:
1. School of Ocean Information Engineering, Jimei University, Xiamen 361021, China 2. Fujlan Electronic Port Co., Ltd., Xiamen 361000, China
Abstract
The effectiveness of the SAR object detection technique based on Convolutional Neural Networks (CNNs) has been widely proven, and it is increasingly used in the recognition of ship targets. Recently, efforts have been made to integrate transformer structures into SAR detectors to achieve improved target localization. However, existing methods rarely design the transformer itself as a detector, failing to fully leverage the long-range modeling advantages of self-attention. Furthermore, there has been limited research into multi-class SAR target detection. To address these limitations, this study proposes a SAR detector named CCDN-DETR, which builds upon the framework of the detection transformer (DETR). To adapt to the multiscale characteristics of SAR data, cross-scale encoders were introduced to facilitate comprehensive information modeling and fusion across different scales. Simultaneously, we optimized the query selection scheme for the input decoder layers, employing IOU loss to assist in initializing object queries more effectively. Additionally, we introduced constrained contrastive denoising training at the decoder layers to enhance the model’s convergence speed and improve the detection of different categories of SAR targets. In the benchmark evaluation on a joint dataset composed of SSDD, HRSID, and SAR-AIRcraft datasets, CCDN-DETR achieves a mean Average Precision (mAP) of 91.9%. Furthermore, it demonstrates significant competitiveness with 83.7% mAP on the multi-class MSAR dataset compared to CNN-based models.
Funder
Youth Program of National Natural Supported by the Science Foundation of China Youth Program of the Natural Science Foundation of Fujian Province of China Xiamen Ocean and Fishery Development Special Fund Project Xiamen Key Laboratory of Marine Intelligent Terminal R&D and Application
Reference51 articles.
1. Li, J., Xu, C., Su, H., Gao, L., and Wang, T. (2022). Deep learning for SAR ship detection: Past, present and future. Remote Sens., 14. 2. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks;Ren;IEEE Trans. Pattern Anal. Mach. Intell.,2017 3. Algorithm/hardware codesign for real-time on-satellite CNN-based ship detection in SAR imagery;Yang;IEEE Trans. Geosci. Remote Sens.,2022 4. Ge, Z., Liu, S., Wang, F., Li, Z., and Sun, J. (2021). Yolox: Exceeding yolo series in 2021. arXiv. 5. Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|