OMNI-CONV: Generalization of the Omnidirectional Distortion-Aware Convolutions
-
Published:2023-01-28
Issue:2
Volume:9
Page:29
-
ISSN:2313-433X
-
Container-title:Journal of Imaging
-
language:en
-
Short-container-title:J. Imaging
Author:
Artizzu Charles-Olivier1ORCID, Allibert Guillaume1, Demonceaux Cédric2
Affiliation:
1. Université Côte d’Azur, CNRS, I3S, 06900 Sophia Antipolis, France 2. ImViA, Université Bourgogne Franche-Comté, 21000 Dijon, France
Abstract
Omnidirectional images have drawn great research attention recently thanks to their great potential and performance in various computer vision tasks. However, processing such a type of image requires an adaptation to take into account spherical distortions. Therefore, it is not trivial to directly extend the conventional convolutional neural networks on omnidirectional images because CNNs were initially developed for perspective images. In this paper, we present a general method to adapt perspective convolutional networks to equirectangular images, forming a novel distortion-aware convolution. Our proposed solution can be regarded as a replacement for the existing convolutional network without requiring any additional training cost. To verify the generalization of our method, we conduct an analysis on three basic vision tasks, i.e., semantic segmentation, optical flow, and monocular depth. The experiments on both virtual and real outdoor scenarios show our adapted spherical models consistently outperform their counterparts.
Funder
French Agence Nationale de la Recherche
Subject
Electrical and Electronic Engineering,Computer Graphics and Computer-Aided Design,Computer Vision and Pattern Recognition,Radiology, Nuclear Medicine and imaging
Reference29 articles.
1. Rituerto, A., Puig, L., and Guerrero, J. (2010, January 27). Comparison of omnidirectional and conventional monocular systems for visual SLAM. Proceedings of the 10th OMNIVIS with Robotics: Science and Systems, Zaragoza, Spain. 2. Won, C., Seok, H., Cui, Z., Pollefeys, M., and Lim, J. (August, January 31). OmniSLAM: Omnidirectional Localization and Dense Mapping for Wide-baseline Multi-camera Systems. Proceedings of the International Conference on Robotics and Automation (ICRA), Paris, France. 3. Artizzu, C.O., Allibert, G., and Demonceaux, C. (2022, January 11–13). Deep Reinforcement Learning with Omnidirectional Images: Application to UAV Navigation in Forests. Proceedings of the 17th International Conference on Control, Automation, Robotics and Vision (ICARCV), Singapore. 4. Chang, A., Dai, A., Funkhouser, T., Halber, M., Niebner, M., Savva, M., Song, S., Zeng, A., and Zhang, Y. (2017, January 10–12). Matterport3D: Learning from RGB-D Data in Indoor Environments. Proceedings of the International Conference on 3D Vision (3DV), Qingdao, China. 5. Armeni, I., Sax, S., Zamir, A.R., and Savarese, S. (2017). Joint 2D-3D-Semantic Data for Indoor Scene Understanding. arXiv.
|
|