IFKD: Implicit field knowledge distillation for single view reconstruction
-
Published:2023
Issue:8
Volume:20
Page:13864-13880
-
ISSN:1551-0018
-
Container-title:Mathematical Biosciences and Engineering
-
language:
-
Short-container-title:MBE
Author:
Wang Jianyuan12, Xu Huanqiang3, Hu Xinrui3, Leng Biao3
Affiliation:
1. School of Intelligence Science and Technology, University of Science and Technology Beijing, Beijing 100083, China 2. Key Laboratory of Intelligent Bionic Unmanned Systems, Ministry of Education, University of Science and Technology Beijing, Beijing 100083, China 3. School of Computer Science and Engineering, Beihang University, Beijing 100191, China
Abstract
<abstract><p>In 3D reconstruction tasks, camera parameter matrix estimation is usually used to present the single view of an object, which is not necessary when mapping the 3D point to 2D image. The single view reconstruction task should care more about the quality of reconstruction instead of the alignment. So in this paper, we propose an implicit field knowledge distillation model (IFKD) to reconstruct 3D objects from the single view. Transformations are performed on 3D points instead of the camera and keep the camera coordinate identified with the world coordinate, so that the extrinsic matrix can be omitted. Besides, a knowledge distillation structure from 3D voxel to the feature vector is established to further refine the feature description of 3D objects. Thus, the details of a 3D model can be better captured by the proposed model. This paper adopts ShapeNet Core dataset to verify the effectiveness of the IFKD model. Experiments show that IFKD has strong advantages in IOU and other core indicators compared with the camera matrix estimation methods, which verifies the feasibility of the new proposed mapping method.</p></abstract>
Publisher
American Institute of Mathematical Sciences (AIMS)
Subject
Applied Mathematics,Computational Mathematics,General Agricultural and Biological Sciences,Modeling and Simulation,General Medicine
Reference30 articles.
1. M. Li, H. Zhang, D2im-net: Learning detail disentangled implicit fields from single images, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2021), 10246–10255. https://doi.org/10.1109/CVPR46437.2021.01011 2. S. Saito, Z. Huang, R. Natsume, S. Morishima, A. Kanazawa, H. Li, Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization, in Proceedings of the IEEE/CVF International Conference on Computer Vision, (2019), 2304–2314. https://doi.org/10.1109/ICCV.2019.00239 3. T. Groueix, M. Fisher, V. G. Kim, B. C. Russell, M. Aubry, A papier-mâché approach to learning 3d surface generation, in Proceedings of the IEEE conference on computer vision and pattern recognition, (2018), 216–224. https://doi.org/10.1109/CVPR.2018.00030 4. C. B. Choy, D. Xu, J. Gwak, K. Chen, S. Savarese, 3d-r2n2: A unified approach for single and multi-view 3d object reconstruction, Eur. Conf. Comput. Vision, 9912 (2016), 628–644. https://doi.org/10.1007/978-3-319-46484-8_38 5. G. Yang, X. Huang, Z. Hao, M. Liu, S. Belongie, B. Hariharan, Pointflow: 3d point cloud generation with continuous normalizing flows, in Proceedings of the IEEE/CVF International Conference on Computer Vision, (2019), 4541–4550. https://doi.org/10.1109/ICCV.2019.00464
|
|