Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Consistency Models (CMs) have significantly accelerated the sampling process
in diffusion models, yielding impressive results in synthesizing
high-resolution images. To explore and extend these advancements to
point-cloud-based 3D shape generation, we propose a novel Multi-scale Latent
Point Consistency Model (MLPCM). Our MLPCM follows a latent diffusion framework
and introduces hierarchical levels of latent representations, ranging from
point-level to super-point levels, each corresponding to a different spatial
resolution. We design a multi-scale latent integration module along with 3D
spatial attention to effectively denoise the point-level latent representations
conditioned on those from multiple super-point levels. Additionally, we propose
a latent consistency model, learned through consistency distillation, that
compresses the prior into a one-step generator. This significantly improves
sampling efficiency while preserving the performance of the original teacher
model. Extensive experiments on standard benchmarks ShapeNet and ShapeNet-Vol
demonstrate that MLPCM achieves a 100x speedup in the generation process, while
surpassing state-of-the-art diffusion models in terms of both shape quality and
diversity.
Authors (3)
Bi'an Du
Wei Hu
Renjie Liao
Submitted
December 27, 2024
Key Contributions
Proposes a novel Multi-scale Latent Point Consistency Model (MLPCM) for 3D shape generation using point clouds. It introduces hierarchical latent representations and a multi-scale integration module to improve denoising, and uses consistency distillation to create a one-step generator, significantly enhancing sampling efficiency.
Business Value
Accelerates the creation of 3D assets for various industries, including gaming, VR/AR, product design, and robotics, by making generative models for 3D shapes much faster and more efficient.