A Novel Dynamic Contextual Feature Fusion Model for Small Object Detection in Satellite Remote-Sensing Images
-
Published:2024-04-18
Issue:4
Volume:15
Page:230
-
ISSN:2078-2489
-
Container-title:Information
-
language:en
-
Short-container-title:Information
Author:
Yang Hongbo12, Qiu Shi1ORCID
Affiliation:
1. Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China 2. University of Chinese Academy of Sciences, Beijing 100049, China
Abstract
Ground objects in satellite images pose unique challenges due to their low resolution, small pixel size, lack of texture features, and dense distribution. Detecting small objects in satellite remote-sensing images is a difficult task. We propose a new detector focusing on contextual information and multi-scale feature fusion. Inspired by the notion that surrounding context information can aid in identifying small objects, we propose a lightweight context convolution block based on dilated convolutions and integrate it into the convolutional neural network (CNN). We integrate dynamic convolution blocks during the feature fusion step to enhance the high-level feature upsampling. An attention mechanism is employed to focus on the salient features of objects. We have conducted a series of experiments to validate the effectiveness of our proposed model. Notably, the proposed model achieved a 3.5% mean average precision (mAP) improvement on the satellite object detection dataset. Another feature of our approach is lightweight design. We employ group convolution to reduce the computational cost in the proposed contextual convolution module. Compared to the baseline model, our method reduces the number of parameters by 30%, computational cost by 34%, and an FPS rate close to the baseline model. We also validate the detection results through a series of visualizations.
Funder
the China Postdoctoral Science Foundation
Reference36 articles.
1. Pereira, F., Burges, C.J., Bottou, L., and Weinberger, K.Q. (2012). Advances in Neural Information Processing Systems 25, Proceedings of the 26th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012, Curran Associates, Inc. 2. Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., and Zitnick, C.L. (2014, January 6–12). Microsoft Coco: Common Objects in Context. Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland. 3. Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27–30). You Only Look Once: Unified, Real-Time Object Detection. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA. 4. Redmon, J., and Farhadi, A. (2018). Yolov3: An Incremental Improvement. arXiv. 5. Bochkovskiy, A., Wang, C.-Y., and Liao, H.-Y.M. (2022). Yolov4: Optimal Speed and Accuracy of Object Detection. arXiv.
|
|