HQ3DAvatar: High-quality Implicit 3D Head Avatar

Author:

Teotia Kartik1ORCID,R Mallikarjun B1ORCID,Pan Xingang2ORCID,Kim Hyeongwoo3ORCID,Garrido Pablo4ORCID,Elgharib Mohamed5ORCID,Theobalt Christian1ORCID

Affiliation:

1. Max Planck Institute for Informatics and Saarland University, Saarbrucken, Germany

2. Max Planck Institute for Informatics, Saarbrucken, Germany and Nanyang Technological University, Singapore, Singapore

3. Imperial College London, London, United Kingdom

4. Flawless AI, Los Angeles, United States of America

5. Max Planck Institute for Informatics, Saarbrucken, Germany

Abstract

Multi-view volumetric rendering techniques have recently shown great potential in modeling and synthesizing high-quality head avatars. A common approach to capture full head dynamic performances is to track the underlying geometry using a mesh-based template or 3D cube-based graphics primitives. While these model-based approaches achieve promising results, they often fail to learn complex geometric details such as the mouth interior, hair, and topological changes over time. This article presents a novel approach to building highly photorealistic digital head avatars. Our method learns a canonical space via an implicit function parameterized by a neural network. It leverages multiresolution hash encoding in the learned feature space, allowing for high quality, faster training, and high-resolution rendering. At test time, our method is driven by a monocular RGB video. Here, an image encoder extracts face-specific features that also condition the learnable canonical space. This encourages deformation-dependent texture variations during training. We also propose a novel optical flow-based loss that ensures correspondences in the learned canonical space, thus encouraging artifact-free and temporally consistent renderings. We show results on challenging facial expressions and show free-viewpoint renderings at interactive real-time rates for a resolution of 480 x 270. Our method outperforms related approaches both visually and numerically. We will release our multiple-identity dataset to encourage further research.

Publisher

Association for Computing Machinery (ACM)

Reference83 articles.

1. 3DAvatarGAN: Bridging domains for personalized editable avatars;Abdal Rameen;CoRR,2023

2. Matthew Amodio David van Dijk Ruth Montgomery Guy Wolf and Smita Krishnaswamy. 2019. Out-of-sample Extrapolation with Neuron Editing. arxiv:q-bio.QM/1805.12198 (2019).

3. RigNeRF: Fully Controllable Neural 3D Portraits

4. High-fidelity facial avatar reconstruction from monocular video with generative priors;Bai Yunpeng;CoRR,2022

5. Alexander W. Bergman, Petr Kellnhofer, Yifan Wang, Eric R. Chan, David B. Lindell, and Gordon Wetzstein. 2022. Generative neural articulated radiance fields. In Conference on Advances in Neural Information Processing Systems.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3