Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 93% Match Research Paper LLM Developers,AI Researchers,Machine Learning Engineers 1 week ago

DiffAdapt: Difficulty-Adaptive Reasoning for Token-Efficient LLM Inference

large-language-models › reasoning
📄 Abstract

Abstract: Recent reasoning Large Language Models (LLMs) demonstrate remarkable problem-solving abilities but often generate long thinking traces whose utility is unclear. Our work aims to improve their efficiency, enabling them to reach high performance without overthinking. First, we analyze the entropy of token probabilities in reasoning traces. Across three models, we observe a consistent U-shaped entropy pattern: high entropy on easy problems despite high accuracy, low entropy on problems with medium difficulty, and high entropy on hard problems reflecting uncertainty. Specifically, we notice 22--25\% entropy reduction from easy to medium difficulty regions, suggesting an {overthinking} phenomenon on easy instances. Building on these insights, we introduce \textbf{DiffAdapt}, a lightweight framework that selects Easy/Normal/Hard inference strategies per question based on their difficulty and reasoning trace entropy. Each inference strategy consists of a fixed prompt, temperature and maximum token length. In contrast to existing efficiency optimization methods, our approach does not fine-tune base LLM but a small probe that classifies LLM's final hidden state, allowing inexpensive adaptation. We comprehensively evaluate our method on five models and eight benchmarks. Our method achieves comparable or improved accuracy while reducing token usage by up to 22.4\%, establishing a practical path toward compute-efficient reasoning.
Authors (4)
Xiang Liu
Xuming Hu
Xiaowen Chu
Eunsol Choi
Submitted
October 22, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Introduces DiffAdapt, a lightweight framework that analyzes reasoning trace entropy to adapt LLM inference strategies (Easy/Normal/Hard) based on problem difficulty. This addresses the 'overthinking' phenomenon observed in LLMs on easy problems, leading to more efficient and accurate reasoning.

Business Value

Enables faster and more cost-effective LLM deployments by optimizing reasoning processes, leading to quicker responses and reduced computational costs for complex problem-solving tasks.