Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Large language models (LLMs) augmented with multi-step reasoning and action
generation abilities have shown promise in leveraging external tools to tackle
complex tasks that require long-horizon planning. However, existing approaches
either rely on implicit planning in the reasoning stage or introduce explicit
planners without systematically addressing how to optimize the planning stage.
As evidence, we observe that under vanilla reinforcement learning (RL),
planning tokens exhibit significantly higher entropy than other action tokens,
revealing uncertain decision points that remain under-optimized. To address
this, we propose DeepPlanner, an end-to-end RL framework that effectively
enhances the planning capabilities of deep research agents. Our approach shapes
token-level advantage with an entropy-based term to allocate larger updates to
high entropy tokens, and selectively upweights sample-level advantages for
planning-intensive rollouts. Extensive experiments across seven deep research
benchmarks demonstrate that DeepPlanner improves planning quality and achieves
state-of-the-art results under a substantially lower training budget.
Authors (9)
Wei Fan
Wenlin Yao
Zheng Li
Feng Yao
Xin Liu
Liang Qiu
+3 more
Submitted
October 14, 2025
Key Contributions
Proposes DeepPlanner, an end-to-end RL framework that enhances planning capabilities in LLM agents. It uses advantage shaping with an entropy-based term to allocate larger updates to high-entropy planning tokens and selectively upweights sample-level advantages for planning-intensive rollouts.
Business Value
Enables the development of more capable AI agents that can autonomously perform complex, multi-step tasks, potentially automating research and development processes.