Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper AI researchers,NLP engineers,Machine learning scientists,Developers of advanced QA systems 1 week ago

LoongRL:Reinforcement Learning for Advanced Reasoning over Long Contexts

large-language-models › reasoning
📄 Abstract

Abstract: Reasoning over long contexts is essential for large language models. While reinforcement learning (RL) enhances short-context reasoning by inducing "Aha" moments in chain-of-thought, the advanced thinking patterns required for long-context reasoning remain largely unexplored, and high-difficulty RL data are scarce. In this paper, we introduce LoongRL, a data-driven RL method for advanced long-context reasoning. Central to LoongRL is KeyChain, a synthesis approach that transforms short multi-hop QA into high-difficulty long-context tasks by inserting UUID chains that hide the true question among large collections of distracting documents. Solving these tasks requires the model to trace the correct chain step-by-step, identify the true question, retrieve relevant facts and reason over them to answer correctly. RL training on KeyChain data induces an emergent plan-retrieve-reason-recheck reasoning pattern that generalizes far beyond training length. Models trained at 16K effectively solve 128K tasks without prohibitive full-length RL rollout costs. On Qwen2.5-7B and 14B, LoongRL substantially improves long-context multi-hop QA accuracy by +23.5% and +21.1% absolute gains. The resulting LoongRL-14B reaches a score of 74.2, rivaling much larger frontier models such as o3-mini (74.5) and DeepSeek-R1 (74.9). It also improves long-context retrieval, passes all 128K needle-in-a-haystack stress tests, and preserves short-context reasoning capabilities.
Authors (7)
Siyuan Wang
Gaokai Zhang
Li Lyna Zhang
Ning Shang
Fan Yang
Dongyao Chen
+1 more
Submitted
October 22, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Introduces LoongRL, a data-driven RL method for advanced long-context reasoning. It uses KeyChain, a synthesis approach to create high-difficulty long-context tasks from short QA, inducing an emergent plan-retrieve-reason-recheck reasoning pattern that generalizes beyond training length.

Business Value

Enables AI systems to process and reason over much larger amounts of information, crucial for tasks like complex document analysis, legal research, and scientific discovery, leading to more powerful knowledge-based applications.