Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Existing Reinforcement Learning from Verifiable Rewards (RLVR) methods, such
as Group Relative Policy Optimization (GRPO), have achieved remarkable progress
in improving the reasoning capabilities of Large Reasoning Models (LRMs).
However, they exhibit limited exploration due to reliance on on-policy rollouts
where confined to the current policy's distribution, resulting in narrow
trajectory diversity. Recent approaches attempt to expand policy coverage by
incorporating trajectories generated from stronger expert models, yet this
reliance increases computational cost and such advaned models are often
inaccessible. To address these issues, we propose In-Context Steered Policy
Optimization (ICPO), a unified framework that leverages the inherent in-context
learning capability of LRMs to provide expert guidance using existing datasets.
ICPO introduces Mixed-Policy GRPO with Implicit Expert Forcing, which expands
exploration beyond the current policy distribution without requiring advanced
LRM trajectories. To further stabilize optimization, ICPO integrates Expert
Region Reject Sampling to filter unreliable off-policy trajectories and
Annealed Expert-Bonus Reward Shaping to balance early expert guidance with
later autonomous improvement. Results demonstrate that ICPO consistently
enhances reinforcement learning performance and training stability on
mathematical reasoning benchmarks, revealing a scalable and effective RLVR
paradigm for LRMs.
Authors (5)
Hsiu-Yuan Huang
Chenming Tang
Weijie Liu
Saiyong Yang
Yunfang Wu
Submitted
October 30, 2025
Key Contributions
ICPO is a unified framework that enhances RLVR by leveraging the in-context learning capability of Large Reasoning Models (LRMs) to provide expert guidance using existing datasets. It introduces Mixed-Policy GRPO with Implicit Expert Forcing, expanding exploration beyond the current policy distribution without requiring external expert models or increasing computational cost significantly.
Business Value
Enables the development of more capable and robust reasoning models, leading to better AI assistants, more reliable content generation, and improved performance in complex decision-making tasks, with potentially lower training costs.