DreamFace: Progressive Generation of Animatable 3D Faces under Text Guidance

Author:

Zhang Longwen12ORCID,Qiu Qiwei12ORCID,Lin Hongyang12ORCID,Zhang Qixuan12ORCID,Shi Cheng1ORCID,Yang Wei3ORCID,Shi Ye1ORCID,Yang Sibei1ORCID,Xu Lan1ORCID,Yu Jingyi1ORCID

Affiliation:

1. ShanghaiTech University, Shanghai, China

2. Deemos Technology, Shanghai, China

3. Huazhong University of Science and Technology, Wuhan, China

Abstract

Emerging Metaverse applications demand accessible, accurate and easy-to-use tools for 3D digital human creations in order to depict different cultures and societies as if in the physical world. Recent large-scale vision-language advances pave the way for novices to conveniently customize 3D content. However, the generated CG-friendly assets still cannot represent the desired facial traits for human characteristics. In this paper, we present Dream-Face, a progressive scheme to generate personalized 3D faces under text guidance. It enables layman users to naturally customize 3D facial assets that are compatible with CG pipelines, with desired shapes, textures and fine-grained animation capabilities. From a text input to describe the facial traits, we first introduce a coarse-to-fine scheme to generate the neutral facial geometry with a unified topology. We employ a selection strategy in the CLIP embedding space to generate coarse geometry, and subsequently optimize both the detailed displacements and normals using Score Distillation Sampling (SDS) from the generic Latent Diffusion Model (LDM). Then, for neutral appearance generation, we introduce a dual-path mechanism, which combines the generic LDM with a novel texture LDM to ensure both the diversity and textural specification in the UV space. We also employ a two-stage optimization to perform SDS in both the latent and image spaces to significantly provide compact priors for fine-grained synthesis. It also enables learning the mapping from the compact latent space into physically-based textures (diffuse albedo, specular intensity, normal maps, etc.). Our generated neutral assets naturally support blendshapes-based facial animations, thanks to the unified geometric topology. We further improve the animation ability with personalized deformation characteristics. To this end, we learn the universal expression prior in a latent space with neutral asset conditioning using the cross-identity hypernetwork, we subsequently train a neural facial tracker from video input space into the pre-trained expression space for personalized fine-grained animation. Extensive qualitative and quantitative experiments validate the effectiveness and generalizability of DreamFace. Notably, DreamFace can generate realistic 3D facial assets with physically-based rendering quality and rich animation ability from video footage, even for fashion icons or exotic characters in cartoons and fiction movies.

Funder

National Key R&D Program of China

NSFC programs

STCSM

SHMEC

Publisher

Association for Computing Machinery (ACM)

Subject

Computer Graphics and Computer-Aided Design

Reference74 articles.

1. A Decoupled 3D Facial Shape Model by Adversarial Training

2. The Digital Emily project

3. Shivangi Aneja Justus Thies Angela Dai and Matthias Nießner. 2022. ClipFace: Text-guided Editing of Textured 3D Morphable Models. In ArXiv preprint arXiv:2212.01406. Shivangi Aneja Justus Thies Angela Dai and Matthias Nießner. 2022. ClipFace: Text-guided Editing of Textured 3D Morphable Models. In ArXiv preprint arXiv:2212.01406.

4. Apple. 2023. ARKit - Face Tracking. https://developer.apple.com/documentation/arkit/arfaceanchor. Apple. 2023. ARKit - Face Tracking. https://developer.apple.com/documentation/arkit/arfaceanchor.

5. AUTOMATIC1111. 2022. stable-diffusion-webui. https://github.com/AUTOMATIC1111/stable-diffusion-webui. AUTOMATIC1111. 2022. stable-diffusion-webui. https://github.com/AUTOMATIC1111/stable-diffusion-webui.

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3