TED-Face: Texture-Enhanced Deep Face Reconstruction in the Wild
Author:
Huang Ying1ORCID, Fang Lin1, Hu Shanfeng2ORCID
Affiliation:
1. Institute of Virtual Reality and Intelligent Systems, Hangzhou Normal University, Hangzhou 311121, China 2. Department of Computer and Information Sciences, Northumbria University, Newcastle-upon-Tyne NE1 8ST, UK
Abstract
We present TED-Face, a new method for recovering high-fidelity 3D facial geometry and appearance with enhanced textures from single-view images. While vision-based face reconstruction has received intensive research in the past decades due to its broad applications, it remains a challenging problem because human eyes are particularly sensitive to numerically minute yet perceptually significant details. Previous methods that seek to minimize reconstruction errors within a low-dimensional face space can suffer from this issue and generate close yet low-fidelity approximations. The loss of high-frequency texture details is a key factor in their process, which we propose to address by learning to recover both dense radiance residuals and sparse facial texture features from a single image, in addition to the variables solved by previous work—shape, appearance, illumination, and camera. We integrate the estimation of all these factors in a single unified deep neural network and train it on several popular face reconstruction datasets. We also introduce two new metrics, visual fidelity (VIF) and structural similarity (SSIM), to compensate for the fact that reconstruction error is not a consistent perceptual metric of quality. On the popular FaceWarehouse facial reconstruction benchmark, our proposed system achieves a VIF score of 0.4802 and an SSIM score of 0.9622, improving over the state-of-the-art Deep3D method by 6.69% and 0.86%, respectively. On the widely used LS3D-300W dataset, we obtain a VIF score of 0.3922 and an SSIM score of 0.9079 for indoor images, and the scores for outdoor images are 0.4100 and 0.9160, respectively, which also represent an improvement over those of Deep3D. These results show that our method is able to recover visually more realistic facial appearance details compared with previous methods.
Funder
National Natural Science Foundation of China Major Project of The National Social Science Fund of China Scientific Research Foundation for Scholars of HZNU Fundamental Research Funds for the Central Universities
Subject
Electrical and Electronic Engineering,Biochemistry,Instrumentation,Atomic and Molecular Physics, and Optics,Analytical Chemistry
Reference61 articles.
1. Prior-Guided Multi-View 3D Head Reconstruction;Wang;IEEE Trans. Multimed.,2022 2. Holistic face processing is induced by shape and texture;Persike;Perception,2013 3. Blanz, V., and Vetter, T. (1999, January 8–13). A morphable model for the synthesis of 3D faces. Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA. 4. Gecer, B., Lattas, A., Ploumpis, S., Deng, J., Papaioannou, A., Moschoglou, S., and Zafeiriou, S. (2020). European Conference on Computer Vision, Springer. 5. Gecer, B., Ploumpis, S., Kotsia, I., and Zafeiriou, S. (2019, January 15–20). Ganfit: Generative adversarial network fitting for high fidelity 3d face reconstruction. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.
|
|