Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 90% Match Research Paper RL Researchers,AI Theorists,Robotics Engineers,Algorithm Designers 2 weeks ago

On the hardness of RL with Lookahead

reinforcement-learning › robotics-rl
📄 Abstract

Abstract: We study reinforcement learning (RL) with transition look-ahead, where the agent may observe which states would be visited upon playing any sequence of $\ell$ actions before deciding its course of action. While such predictive information can drastically improve the achievable performance, we show that using this information optimally comes at a potentially prohibitive computational cost. Specifically, we prove that optimal planning with one-step look-ahead ($\ell=1$) can be solved in polynomial time through a novel linear programming formulation. In contrast, for $\ell \geq 2$, the problem becomes NP-hard. Our results delineate a precise boundary between tractable and intractable cases for the problem of planning with transition look-ahead in reinforcement learning.
Authors (5)
Corentin Pla
Hugo Richard
Marc Abeille
Nadav Merlis
Vianney Perchet
Submitted
October 22, 2025
arXiv Category
stat.ML
arXiv PDF

Key Contributions

Establishes a precise boundary for the computational hardness of reinforcement learning with transition lookahead. It proves that one-step lookahead is polynomial-time solvable via linear programming, while two or more steps become NP-hard, highlighting the significant computational cost of advanced predictive information.

Business Value

Provides crucial theoretical insights for designing practical RL systems, guiding developers on the feasibility of incorporating lookahead mechanisms based on computational constraints.