One-Stage Small Object Detection Using Super-Resolved Feature Map for Edge Devices
-
Published:2024-01-18
Issue:2
Volume:13
Page:409
-
ISSN:2079-9292
-
Container-title:Electronics
-
language:en
-
Short-container-title:Electronics
Author:
Huynh Xuan Nghia1ORCID, Jung Gu Beom1ORCID, Suhr Jae Kyu1ORCID
Affiliation:
1. Department of Intelligent Mechatronics Engineering, Sejong University, Seoul 05006, Republic of Korea
Abstract
Despite the achievements of deep neural-network-based object detection, detecting small objects in low-resolution images remains a challenging task due to limited information. A possible solution to alleviate the issue involves integrating super-resolution (SR) techniques into object detectors, particularly enhancing feature maps for small-sized objects. This paper explores the impact of high-resolution super-resolved feature maps generated by SR techniques, especially for a one-stage detector that demonstrates a good compromise between detection accuracy and computational efficiency. Firstly, this paper suggests the integration of an SR module named feature texture transfer (FTT) into the one-stage detector, YOLOv4. Feature maps from the backbone and the neck of vanilla YOLOv4 are combined to build a super-resolved feature map for small-sized object detection. Secondly, it proposes a novel SR module with more impressive performance and slightly lower computation demand than the FTT. The proposed SR module utilizes three input feature maps with different resolutions to generate a super-resolved feature map for small-sized object detection. Lastly, it introduces a simplified version of an SR module that maintains similar performance while using only half the computation of the FTT. This attentively simplified module can be effectively used for real-time embedded systems. Experimental results demonstrate that the proposed approach substantially enhances the detection performance of small-sized objects on two benchmark datasets, including a self-built surveillance dataset and the VisDrone2019 dataset. In addition, this paper employs the proposed approach on an embedded system with a Qualcomm QCS610 and demonstrates its feasibility for real-time operation on edge devices.
Funder
National Research Foundation of Korea
Reference45 articles.
1. Object Detection in 20 Years: A Survey;Zou;Proc. IEEE Inst. Electr. Electron. Eng.,2023 2. Girshick, R. (2015, January 7–13). Fast R-CNN. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile. 3. Ren, S., He, K., Girshick, R., and Sun, J. (2015, January 7–10). Faster r-cnn: Towards real-time object detection with region proposal networks. Proceedings of the 28th Advances in Neural Information Processing Systems (NIPS’15), Montreal, QC, Canada. 4. Lin, T.Y., Goyal, P., Girshick, R., He, K., and Dollar, P. (2017, January 22–29). Focal Loss for Dense Object Detection. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy. 5. Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016, January 27–30). You Only Look Once: Unified, Real-Time Object Detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.
|
|