Affiliation:
1. Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
Abstract
3D (Three-Dimensional) scene inpainting aims to remove objects from scenes and generate visually plausible regions to fill the hollows. Leveraging the foundation of NeRF (Neural Radiance Field), considerable advancements have been achieved in the realm of 3D scene inpainting. However, prevalent issues persist: primarily, the presence of inconsistent 3D details across different viewpoints and occlusion losses of real background details in inpainted regions. This paper presents a NeRF-based inpainting approach using uncertainty estimation that formulates mask and uncertainty branches for consistency enhancement. In the initial training, the mask branch learns a 3D-consistent representation from inaccurate input masks, and after background rendering, the background regions can be fully exposed to the views. The uncertainty branch learns the visibility of spatial points by modeling them as Gaussian distributions, generating variances to identify regions to be inpainted. During the inpainting training phase, the uncertainty branch measures 3D consistency in the inpainted views and calculates the confidence from the variance as dynamic weights, which are used to balance the color and adversarial losses to achieve 3D-consistent inpainting with both the structure and texture. The results were evaluated on datasets such as Spin-NeRF and NeRF-Object-Removal. The proposed approach outperformed the baselines in inpainting metrics of LPIPS and FID, and preserved more spatial details from real backgrounds in multi-scene settings, thus achieving 3D-consistent restoration.
Funder
National Natural Science Foundation of China
Yunnan Provincial Science and Technology Plan Project
Faculty of Information Engineering and Automation, Kunming University of Science and Technology
Reference56 articles.
1. Nerf: Representing scenes as neural radiance fields for view synthesis;Mildenhall;Commun. ACM,2021
2. Display of surfaces from volume data;Levoy;IEEE Comput. Graph. Appl.,1988
3. Instant neural graphics primitives with a multiresolution hash encoding;Evans;ACM Trans. Graph. (ToG),2022
4. Fridovich-Keil, S., Yu, A., Tancik, M., Chen, Q., Recht, B., and Kanazawa, A. (2022, January 18–24). Plenoxels: Radiance fields without neural networks. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.
5. Chen, A., Xu, Z., Geiger, A., Yu, J., and Su, H. (2022, January 23–27). Tensorf: Tensorial radiance fields. Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献