Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

3DAvatarGAN: Bridging Domains for Personalized Editable Avatars (2301.02700v2)

Published 6 Jan 2023 in cs.CV and cs.GR

Abstract: Modern 3D-GANs synthesize geometry and texture by training on large-scale datasets with a consistent structure. Training such models on stylized, artistic data, with often unknown, highly variable geometry, and camera information has not yet been shown possible. Can we train a 3D GAN on such artistic data, while maintaining multi-view consistency and texture quality? To this end, we propose an adaptation framework, where the source domain is a pre-trained 3D-GAN, while the target domain is a 2D-GAN trained on artistic datasets. We then distill the knowledge from a 2D generator to the source 3D generator. To do that, we first propose an optimization-based method to align the distributions of camera parameters across domains. Second, we propose regularizations necessary to learn high-quality texture, while avoiding degenerate geometric solutions, such as flat shapes. Third, we show a deformation-based technique for modeling exaggerated geometry of artistic domains, enabling -- as a byproduct -- personalized geometric editing. Finally, we propose a novel inversion method for 3D-GANs linking the latent spaces of the source and the target domains. Our contributions -- for the first time -- allow for the generation, editing, and animation of personalized artistic 3D avatars on artistic datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Rameen Abdal (15 papers)
  2. Hsin-Ying Lee (60 papers)
  3. Peihao Zhu (15 papers)
  4. Menglei Chai (37 papers)
  5. Aliaksandr Siarohin (58 papers)
  6. Peter Wonka (130 papers)
  7. Sergey Tulyakov (108 papers)
Citations (43)

Summary

We haven't generated a summary for this paper yet.