Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Existing end-to-end autonomous driving (AD) algorithms typically follow the
Imitation Learning (IL) paradigm, which faces challenges such as causal
confusion and an open-loop gap. In this work, we propose RAD, a 3DGS-based
closed-loop Reinforcement Learning (RL) framework for end-to-end Autonomous
Driving. By leveraging 3DGS techniques, we construct a photorealistic digital
replica of the real physical world, enabling the AD policy to extensively
explore the state space and learn to handle out-of-distribution scenarios
through large-scale trial and error. To enhance safety, we design specialized
rewards to guide the policy in effectively responding to safety-critical events
and understanding real-world causal relationships. To better align with human
driving behavior, we incorporate IL into RL training as a regularization term.
We introduce a closed-loop evaluation benchmark consisting of diverse,
previously unseen 3DGS environments. Compared to IL-based methods, RAD achieves
stronger performance in most closed-loop metrics, particularly exhibiting a 3x
lower collision rate. Abundant closed-loop results are presented in the
supplementary material. Code is available at https://github.com/hustvl/RAD for
facilitating future research.
Authors (14)
Hao Gao
Shaoyu Chen
Bo Jiang
Bencheng Liao
Yiang Shi
Xiaoyang Guo
+8 more
Submitted
February 18, 2025
Key Contributions
Introduces RAD, a closed-loop Reinforcement Learning framework for end-to-end autonomous driving, leveraging 3D Gaussian Splatting (3DGS) for photorealistic simulation. This allows for extensive state space exploration and learning to handle out-of-distribution scenarios, while specialized rewards enhance safety and IL regularization aligns with human behavior.
Business Value
Accelerates the development and validation of safer and more robust autonomous driving systems by enabling large-scale, realistic simulation-based training.