Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
This paper proposes a novel framework to directly address the Sim2Real performance gap in Reinforcement Learning by adapting simulator parameters based on real-world performance. It frames this as a bi-level RL problem where an outer-level RL agent tunes the simulation model and reward parameters to maximize real-world policy performance, overcoming the limitation that current methods optimize indirect proxies.
Enables more reliable and efficient development of robotic systems and autonomous agents by reducing the need for extensive real-world testing and fine-tuning.