Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Masked diffusion models (MDM) are powerful generative models for discrete
data that generate samples by progressively unmasking tokens in a sequence.
Each token can take one of two states: masked or unmasked. We observe that
token sequences often remain unchanged between consecutive sampling steps;
consequently, the model repeatedly processes identical inputs, leading to
redundant computation. To address this inefficiency, we propose the Partial
masking scheme (Prime), which augments MDM by allowing tokens to take
intermediate states interpolated between the masked and unmasked states. This
design enables the model to make predictions based on partially observed token
information, and facilitates a fine-grained denoising process. We derive a
variational training objective and introduce a simple architectural design to
accommodate intermediate-state inputs. Our method demonstrates superior
performance across a diverse set of generative modeling tasks. On text data, it
achieves a perplexity of 15.36 on OpenWebText, outperforming previous MDM
(21.52), autoregressive models (17.54), and their hybrid variants (17.58),
without relying on an autoregressive formulation. On image data, it attains
competitive FID scores of 3.26 on CIFAR-10 and 6.98 on ImageNet-32, comparable
to leading continuous generative models.
Authors (5)
Chen-Hao Chao
Wei-Fang Sun
Hanwen Liang
Chun-Yi Lee
Rahul G. Krishnan
Key Contributions
This paper proposes a Partial Masking scheme (Prime) for discrete diffusion models, enhancing Masked Diffusion Models (MDM) by allowing tokens to take intermediate states. This 'Prime' scheme enables finer-grained denoising and reduces redundant computation by avoiding repeated processing of identical inputs. The method demonstrates superior performance across diverse generative modeling tasks, particularly on text data.
Business Value
More efficient and effective generation of discrete data, such as text, can lead to improved AI writing assistants, code generation tools, and more sophisticated content creation platforms.