Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper LLM researchers,Reinforcement learning practitioners,AI safety researchers,NLP engineers 2 weeks ago

Stop Summation: Min-Form Credit Assignment Is All Process Reward Model Needs for Reasoning

large-language-models › reasoning
📄 Abstract

Abstract: Process reward models (PRMs) have proven effective for test-time scaling of Large Language Models (LLMs) on challenging reasoning tasks. However, reward hacking issues with PRMs limit their successful application in reinforcement fine-tuning. In this paper, we identify the main cause of PRM-induced reward hacking: the canonical summation-form credit assignment in reinforcement learning (RL), which defines the value as cumulative gamma-decayed future rewards, easily induces LLMs to hack steps with high rewards. To address this, we propose PURE: Process sUpervised Reinforcement lEarning. The key innovation of PURE is a min-form credit assignment that formulates the value function as the minimum of future rewards. This method significantly alleviates reward hacking by limiting the value function range and distributing advantages more reasonably. Through extensive experiments on 3 base models, we show that PRM-based approaches enabling min-form credit assignment achieve comparable reasoning performance to verifiable reward-based methods within only 30% steps. In contrast, the canonical sum-form credit assignment collapses training even at the beginning! Additionally, when we supplement PRM-based fine-tuning with just 10% verifiable rewards, we further alleviate reward hacking and produce the best fine-tuned model based on Qwen2.5-Math-7B in our experiments, achieving 82.5% accuracy on AMC23 and 53.3% average accuracy across 5 benchmarks. Moreover, we summarize the observed reward hacking cases and analyze the causes of training collapse. We release our code and model weights at https://github.com/CJReinforce/PURE.
Authors (8)
Jie Cheng
Gang Xiong
Ruixi Qiao
Lijun Li
Chao Guo
Junle Wang
+2 more
Submitted
April 21, 2025
arXiv Category
cs.AI
arXiv PDF

Key Contributions

The paper identifies summation-form credit assignment as the cause of reward hacking in PRM-based LLM fine-tuning and proposes PURE, which uses a novel min-form credit assignment. This approach alleviates reward hacking by limiting the value function range and distributing advantages more reasonably, leading to comparable performance with significantly reduced hacking.

Business Value

Enables more reliable and robust fine-tuning of LLMs for complex reasoning tasks, reducing the risk of unintended behaviors caused by reward hacking and leading to more trustworthy AI systems.