Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Reinforcement Learning (RL) can mitigate the causal confusion and
distribution shift inherent to imitation learning (IL). However, applying RL to
end-to-end autonomous driving (E2E-AD) remains an open problem for its training
difficulty, and IL is still the mainstream paradigm in both academia and
industry. Recently Model-based Reinforcement Learning (MBRL) have demonstrated
promising results in neural planning; however, these methods typically require
privileged information as input rather than raw sensor data. We fill this gap
by designing Raw2Drive, a dual-stream MBRL approach. Initially, we efficiently
train an auxiliary privileged world model paired with a neural planner that
uses privileged information as input. Subsequently, we introduce a raw sensor
world model trained via our proposed Guidance Mechanism, which ensures
consistency between the raw sensor world model and the privileged world model
during rollouts. Finally, the raw sensor world model combines the prior
knowledge embedded in the heads of the privileged world model to effectively
guide the training of the raw sensor policy. Raw2Drive is so far the only RL
based end-to-end method on CARLA Leaderboard 2.0, and Bench2Drive and it
achieves state-of-the-art performance.
Authors (6)
Zhenjie Yang
Xiaosong Jia
Qifeng Li
Xue Yang
Maoqing Yao
Junchi Yan
Key Contributions
Raw2Drive proposes a dual-stream MBRL approach for end-to-end autonomous driving that effectively trains a raw sensor world model using privileged information. It employs an auxiliary privileged world model and a neural planner, and introduces a Guidance Mechanism to ensure consistency between the raw and privileged world models during rollouts.
Business Value
Paves the way for more robust and adaptable autonomous driving systems by leveraging RL and world models trained on realistic sensor inputs, potentially improving safety and performance.