Abstract
We present a learning-based method for building driving-signal aware full-body avatars. Our model is a conditional variational autoencoder that can be animated with incomplete driving signals, such as human pose and facial keypoints, and produces a high-quality representation of human geometry and view-dependent appearance. The core intuition behind our method is that better drivability and generalization can be achieved by disentangling the driving signals and remaining generative factors, which are not available during animation. To this end, we explicitly account for information deficiency in the driving signal by introducing a latent space that exclusively captures the remaining information, thus enabling the imputation of the missing factors required during full-body animation, while remaining faithful to the driving signal. We also propose a learnable localized compression for the driving signal which promotes better generalization, and helps minimize the influence of global chance-correlations often found in real datasets. For a given driving signal, the resulting variational model produces a compact space of uncertainty for missing factors that allows for an imputation strategy best suited to a particular application. We demonstrate the efficacy of our approach on the challenging problem of full-body animation for virtual telepresence with driving signals acquired from minimal sensors placed in the environment and mounted on a VR-headset.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Graphics and Computer-Aided Design
Cited by
37 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. LayGA: Layered Gaussian Avatars for Animatable Clothing Transfer;Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers '24;2024-07-13
2. MoRF: Mobile Realistic Fullbody Avatars from a Monocular Video;2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV);2024-01-03
3. A Two-Step Approach for Interactive Animatable Avatars;Advances in Computer Graphics;2023-12-29
4. Single-Image 3D Human Digitization with Shape-guided Diffusion;SIGGRAPH Asia 2023 Conference Papers;2023-12-10
5. Drivable Avatar Clothing: Faithful Full-Body Telepresence with Dynamic Clothing Driven by Sparse RGB-D Input;SIGGRAPH Asia 2023 Conference Papers;2023-12-10