Affiliation:
1. CAS and University of Chinese Academy of Sciences
2. Cardiff University
3. City University of Hong Kong
Abstract
Sketches, which are simple and concise, have been used in recent deep image synthesis methods to allow intuitive generation and editing of facial images. However, it is nontrivial to extend such methods to video editing due to various challenges, ranging from appropriate manipulation propagation and fusion of multiple editing operations to ensure temporal coherence and visual quality. To address these issues, we propose a novel sketch-based facial video editing framework, in which we represent editing manipulations in latent space and propose specific propagation and fusion modules to generate high-quality video editing results based on StyleGAN3. Specifically, we first design an optimization approach to represent sketch editing manipulations by editing vectors, which are propagated to the whole video sequence using a proper strategy to cope with different editing needs. Specifically, input editing operations are classified into two categories: temporally consistent editing and temporally variant editing. The former (e.g., change of face shape) is applied to the whole video sequence directly, while the latter (e.g., change of facial expression or dynamics) is propagated with the guidance of expression or only affects adjacent frames in a given time window. Since users often perform different editing operations in multiple frames, we further present a region-aware fusion approach to fuse diverse editing effects. Our method supports video editing on facial structure and expression movement by sketch, which cannot be achieved by previous works. Both qualitative and quantitative evaluations show the superior editing ability of our system to existing and alternative solutions.
Funder
the National Natural Science Foundation of China
Royal Society Newton Advanced Fellowship
the Beijing Municipal Natural Science Foundation for Distinguished Young Scholars
the Youth Innovation Promotion Association CAS
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Graphics and Computer-Aided Design
Reference56 articles.
1. StyleFlow: Attribute-conditioned Exploration of StyleGAN-Generated Images using Conditional Continuous Normalizing Flows
2. Only a matter of style
3. Yuval Alaluf , Omer Tov , Ron Mokady , Rinon Gal , and Amit H . Bermano . 2021 b. HyperStyle: Style GAN Inversion with HyperNetworks for Real Image Editing. CoRR abs/2111.15666 (2021). Yuval Alaluf, Omer Tov, Ron Mokady, Rinon Gal, and Amit H. Bermano. 2021b. HyperStyle: StyleGAN Inversion with HyperNetworks for Real Image Editing. CoRR abs/2111.15666 (2021).
4. Flow-based video synthesis and editing
5. Mikolaj Binkowski Danica J. Sutherland Michael Arbel and Arthur Gretton. 2018. Demystifying MMD GANs. In ICLR. Mikolaj Binkowski Danica J. Sutherland Michael Arbel and Arthur Gretton. 2018. Demystifying MMD GANs. In ICLR.
Cited by
14 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献