Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Diffusion-based large language models (dLLMs) have exhibited substantial
potential for parallel text generation, which may enable more efficient
generation compared to autoregressive models. However, current dLLMs suffer
from fixed generation lengths, which indicates the generation lengths of dLLMs
have to be determined before decoding as a hyper-parameter, leading to issues
in efficiency and flexibility. To solve these problems, in this work, we
propose to train a diffusion LLM with native variable generation lengths,
abbreviated as dLLM-Var. Concretely, we aim to train a model to accurately
predict the [EOS] token in the generated text, which makes a dLLM be able to
natively infer in a block diffusion manner, while still maintaining the ability
of global bi-directional (full) attention and high parallelism. Experiments on
standard benchmarks demonstrate that our method achieves a 30.1x speedup over
traditional dLLM inference paradigms and a 2.4x speedup relative to
autoregressive models such as Qwen and Llama. Our method achieves higher
accuracy and faster inference, elevating dLLMs beyond mere academic novelty and
supporting their practical use in real-world applications. Codes and models
have been released.
Authors (7)
Yicun Yang
Cong Wang
Shaobo Wang
Zichen Wen
Biqing Qi
Hanlin Xu
+1 more
Submitted
October 28, 2025
Key Contributions
Proposes dLLM-Var, a diffusion-based LLM capable of native variable generation lengths by accurately predicting the [EOS] token. This enables block diffusion inference, achieving significant speedups over traditional dLLM inference and autoregressive models while maintaining parallelism and global attention.
Business Value
Enables faster and more flexible text generation, improving efficiency for applications like content creation, summarization, and dialogue systems.