Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Recent reasoning Large Language Models (LLMs) demonstrate remarkable
problem-solving abilities but often generate long thinking traces whose utility
is unclear. Our work aims to improve their efficiency, enabling them to reach
high performance without overthinking. First, we analyze the entropy of token
probabilities in reasoning traces. Across three models, we observe a consistent
U-shaped entropy pattern: high entropy on easy problems despite high accuracy,
low entropy on problems with medium difficulty, and high entropy on hard
problems reflecting uncertainty. Specifically, we notice 22--25\% entropy
reduction from easy to medium difficulty regions, suggesting an {overthinking}
phenomenon on easy instances. Building on these insights, we introduce
\textbf{DiffAdapt}, a lightweight framework that selects Easy/Normal/Hard
inference strategies per question based on their difficulty and reasoning trace
entropy. Each inference strategy consists of a fixed prompt, temperature and
maximum token length. In contrast to existing efficiency optimization methods,
our approach does not fine-tune base LLM but a small probe that classifies
LLM's final hidden state, allowing inexpensive adaptation. We comprehensively
evaluate our method on five models and eight benchmarks. Our method achieves
comparable or improved accuracy while reducing token usage by up to 22.4\%,
establishing a practical path toward compute-efficient reasoning.
Authors (4)
Xiang Liu
Xuming Hu
Xiaowen Chu
Eunsol Choi
Submitted
October 22, 2025
Key Contributions
Introduces DiffAdapt, a lightweight framework that analyzes reasoning trace entropy to adapt LLM inference strategies (Easy/Normal/Hard) based on problem difficulty. This addresses the 'overthinking' phenomenon observed in LLMs on easy problems, leading to more efficient and accurate reasoning.
Business Value
Enables faster and more cost-effective LLM deployments by optimizing reasoning processes, leading to quicker responses and reduced computational costs for complex problem-solving tasks.