Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: We present a novel, zero-shot pipeline for creating hyperrealistic,
identity-preserving 3D avatars from a few unstructured phone images. Existing
methods face several challenges: single-view approaches suffer from geometric
inconsistencies and hallucinations, degrading identity preservation, while
models trained on synthetic data fail to capture high-frequency details like
skin wrinkles and fine hair, limiting realism. Our method introduces two key
contributions: (1) a generative canonicalization module that processes multiple
unstructured views into a standardized, consistent representation, and (2) a
transformer-based model trained on a new, large-scale dataset of high-fidelity
Gaussian splatting avatars derived from dome captures of real people. This
"Capture, Canonicalize, Splat" pipeline produces static quarter-body avatars
with compelling realism and robust identity preservation from unstructured
photos.
Authors (17)
Emanuel Garbin
Guy Adam
Oded Krams
Zohar Barzelay
Eran Guendelman
Michael Schwarz
+11 more
Submitted
October 15, 2025
Key Contributions
Introduces a novel, zero-shot pipeline for creating hyperrealistic, identity-preserving 3D avatars from unstructured phone images. It features a generative canonicalization module for consistent view processing and a transformer model trained on a new large-scale dataset of Gaussian splatting avatars, overcoming limitations of single-view methods and synthetic data training.
Business Value
Enables the creation of highly realistic and personalized 3D avatars from readily available phone images, significantly lowering the barrier to entry for applications in the metaverse, gaming, and virtual communication.