Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation
Author:
Lo Chen-Chou1ORCID, Vandewalle Patrick1ORCID
Affiliation:
1. Processing Speech and Images (PSI), Department of Electrical Engineering (ESAT), KU Leuven, 3001 Leuven, Belgium
Abstract
Radar data can provide additional depth information for monocular depth estimation. It provides a cost-effective solution and is robust in various weather conditions, particularly when compared with lidar. Given the sparse and limited vertical field of view of radar signals, existing methods employ either a vertical extension of radar points or the training of a preprocessing neural network to extend sparse radar points under lidar supervision. In this work, we present a novel radar expansion technique inspired by the joint bilateral filter, tailored for radar-guided monocular depth estimation. Our approach is motivated by the synergy of spatial and range kernels within the joint bilateral filter. Unlike traditional methods that assign a weighted average of nearby pixels to the current pixel, we expand sparse radar points by calculating a confidence score based on the values of spatial and range kernels. Additionally, we propose the use of a range-aware window size for radar expansion instead of a fixed window size in the image plane. Our proposed method effectively increases the number of radar points from an average of 39 points in a raw radar frame to an average of 100 K points. Notably, the expanded radar exhibits fewer intrinsic errors when compared with raw radar and previous methodologies. To validate our approach, we assess our proposed depth estimation model on the nuScenes dataset. Comparative evaluations with existing radar-guided depth estimation models demonstrate its state-of-the-art performance.
Funder
KU Leuven-Taiwan MOE Scholarship Internal Funds KU Leuven
Reference43 articles.
1. Chang, J., and Chen, Y. (2018, January 18–23). Pyramid Stereo Matching Network. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA. 2. Zhang, F., Prisacariu, V., Yang, R., and Torr, P. (2019, January 15–20). GA-Net: Guided Aggregation Net for End-To-End Stereo Matching. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA. 3. Eigen, D., Puhrsch, C., and Fergus, R. (2014, January 8–13). Depth Map Prediction from a Single Image using a Multi-Scale Deep Network. Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, Canada. 4. Eigen, D., and Fergus, R. (2015, January 7–13). Predicting Depth, Surface Normals, and Semantic Labels with a Common Multi-scale Convolutional Architecture. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Santiago, Chile. 5. Fu, H., Gong, M., Wang, C., Batmanghelich, K., and Tao, D. (2018, January 18–23). Deep Ordinal Regression Network for Monocular Depth Estimation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.
|
|