Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 92% Match Research Paper ML Researchers,Deep Learning Engineers,AI Infrastructure Specialists 1 week ago

Through the River: Understanding the Benefit of Schedule-Free Methods for Language Model Training

large-language-models › training-methods
📄 Abstract

Abstract: As both model and dataset sizes continue to scale rapidly, conventional pretraining strategies with fixed compute budgets-such as cosine learning rate schedules-are increasingly inadequate for large-scale training. Recent alternatives, including warmup-stable-decay (WSD) schedules and weight averaging, offer greater flexibility. However, WSD relies on explicit decay phases to track progress, while weight averaging addresses this limitation at the cost of additional memory. In search of a more principled and scalable alternative, we revisit the Schedule-Free (SF) method [Defazio et al., 2024], which has shown strong empirical performance across diverse settings. We show that SF-AdamW effectively navigates the "river" structure of the loss landscape without decay phases or auxiliary averaging, making it particularly suitable for continuously scaling training workloads. To understand this behavior, we conduct a theoretical and empirical analysis of SF dynamics, revealing that it implicitly performs weight averaging without memory overhead. Guided by this analysis, we propose a refined variant of SF that improves robustness to momentum and performs better under large batch sizes, addressing key limitations of the original method. Together, these results establish SF as a practical, scalable, and theoretically grounded approach for language model training.
Authors (4)
Minhak Song
Beomhan Baek
Kwangjun Ahn
Chulhee Yun
Submitted
July 14, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Provides a theoretical and empirical analysis of the Schedule-Free (SF) method, demonstrating that SF-AdamW effectively navigates the loss landscape without requiring explicit decay phases or auxiliary weight averaging. This makes it particularly suitable for continuously scaling training workloads, offering a more principled and scalable alternative to conventional pretraining strategies.

Business Value

Enables more efficient and stable training of very large AI models, reducing training time and computational costs. This can accelerate the development and deployment of advanced AI systems.