Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Discrete diffusion models offer a promising alternative to autoregressive
generation through parallel decoding, but they suffer from a sampling wall:
once categorical sampling occurs, rich distributional information collapses
into one-hot vectors and cannot be propagated across steps, forcing subsequent
steps to operate with limited information. To mitigate this problem, we
introduce Loopholing, a novel and simple mechanism that preserves this
information via a deterministic latent pathway, leading to Loopholing Discrete
Diffusion Models (LDDMs). Trained efficiently with a self-conditioning
strategy, LDDMs achieve substantial gains-reducing generative perplexity by up
to 61% over prior baselines, closing (and in some cases surpassing) the gap
with autoregressive models, and producing more coherent text. Applied to
reasoning tasks, LDDMs also improve performance on arithmetic benchmarks such
as Countdown and Game of 24. These results also indicate that loopholing
mitigates idle steps and oscillations, providing a scalable path toward
high-quality non-autoregressive text generation.
Authors (5)
Mingyu Jo
Jaesik Yoon
Justin Deschenaux
Caglar Gulcehre
Sungjin Ahn
Submitted
October 22, 2025
Key Contributions
Introduces Loopholing, a novel mechanism for discrete diffusion models that preserves distributional information via a deterministic latent pathway, overcoming the 'sampling wall' problem. This leads to Loopholing Discrete Diffusion Models (LDDMs) that achieve significant gains in generative perplexity and improve performance on reasoning tasks.
Business Value
Enables faster and more coherent text generation, potentially improving applications like chatbots, content creation, and AI assistants by overcoming limitations of current diffusion models.