Fisheye Object Detection with Visual Prompting-Aided Fine-Tuning
-
Published:2024-06-07
Issue:12
Volume:16
Page:2054
-
ISSN:2072-4292
-
Container-title:Remote Sensing
-
language:en
-
Short-container-title:Remote Sensing
Author:
Jeon Minwoo1ORCID, Park Gyeong-Moon1ORCID, Hwang Hyoseok1ORCID
Affiliation:
1. College of Software, Kyunghee University, Yongin 17104, Republic of Korea
Abstract
Fisheye cameras play a crucial role in various fields by offering a wide field of view, enabling the capture of expansive areas within a single frame. Nonetheless, the radial distortion characteristics of fisheye lenses lead to notable shape deformation, particularly at the edges of the image, posing a significant challenge for accurate object detection. In this paper, we introduce a novel method, ‘VP-aided fine-tuning’, which harnesses the strengths of the pretraining–fine-tuning paradigm augmented by visual prompting (VP) to bridge the domain gap between undistorted standard datasets and distorted fisheye image datasets. Our approach involves two key elements: the use of VPs to effectively adapt a pretrained model to the fisheye domain, and a detailed 24-point regression of objects to fit the unique distortions of fisheye images. This 24-point regression accurately defines the object boundaries and substantially reduces the impact of environmental noise. The proposed method was evaluated against existing object detection frameworks on fisheye images, demonstrating superior performance and robustness. Experimental results also showed performance improvements with the application of VP, regardless of the variety of fine-tuning method applied.
Funder
Institute of Information & communications Technology Planning & Evaluation Convergence security core talent training business support program
Reference45 articles.
1. Approximate model of fisheye camera based on the optical refraction;Zhu;Multimed. Tools Appl.,2014 2. Choi, K.H., Kim, Y., and Kim, C. (2019). Analysis of Fish-Eye Lens Camera Self-Calibration. Sensors, 19. 3. Yogamani, S., Hughes, C., Horgan, J., Sistu, G., Varley, P., O’Dea, D., Uricár, M., Milz, S., Simon, M., and Amende, K. (2019, January 27–28). Woodscape: A multi-task, multi-camera fisheye dataset for autonomous driving. Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea. 4. Panoramic SLAM from a multiple fisheye camera rig;Ji;ISPRS J. Photogramm. Remote Sens.,2020 5. Xiong, Y., and Turkowski, K. (1997, January 17–19). Creating image-based VR using a self-calibrating fisheye lens. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA.
|
|