Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: As both model and dataset sizes continue to scale rapidly, conventional
pretraining strategies with fixed compute budgets-such as cosine learning rate
schedules-are increasingly inadequate for large-scale training. Recent
alternatives, including warmup-stable-decay (WSD) schedules and weight
averaging, offer greater flexibility. However, WSD relies on explicit decay
phases to track progress, while weight averaging addresses this limitation at
the cost of additional memory. In search of a more principled and scalable
alternative, we revisit the Schedule-Free (SF) method [Defazio et al., 2024],
which has shown strong empirical performance across diverse settings. We show
that SF-AdamW effectively navigates the "river" structure of the loss landscape
without decay phases or auxiliary averaging, making it particularly suitable
for continuously scaling training workloads. To understand this behavior, we
conduct a theoretical and empirical analysis of SF dynamics, revealing that it
implicitly performs weight averaging without memory overhead. Guided by this
analysis, we propose a refined variant of SF that improves robustness to
momentum and performs better under large batch sizes, addressing key
limitations of the original method. Together, these results establish SF as a
practical, scalable, and theoretically grounded approach for language model
training.
Authors (4)
Minhak Song
Beomhan Baek
Kwangjun Ahn
Chulhee Yun
Key Contributions
Provides a theoretical and empirical analysis of the Schedule-Free (SF) method, demonstrating that SF-AdamW effectively navigates the loss landscape without requiring explicit decay phases or auxiliary weight averaging. This makes it particularly suitable for continuously scaling training workloads, offering a more principled and scalable alternative to conventional pretraining strategies.
Business Value
Enables more efficient and stable training of very large AI models, reducing training time and computational costs. This can accelerate the development and deployment of advanced AI systems.