Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Recently, Vision-Language-Action models (VLA) have advanced robot imitation
learning, but high data collection costs and limited demonstrations hinder
generalization and current imitation learning methods struggle in
out-of-distribution scenarios, especially for long-horizon tasks. A key
challenge is how to mitigate compounding errors in imitation learning, which
lead to cascading failures over extended trajectories. To address these
challenges, we propose the Diffusion Trajectory-guided Policy (DTP) framework,
which generates 2D trajectories through a diffusion model to guide policy
learning for long-horizon tasks. By leveraging task-relevant trajectories, DTP
provides trajectory-level guidance to reduce error accumulation. Our two-stage
approach first trains a generative vision-language model to create
diffusion-based trajectories, then refines the imitation policy using them.
Experiments on the CALVIN benchmark show that DTP outperforms state-of-the-art
baselines by 25% in success rate, starting from scratch without external
pretraining. Moreover, DTP significantly improves real-world robot performance.
Key Contributions
The Diffusion Trajectory-guided Policy (DTP) framework addresses compounding errors in imitation learning for long-horizon robot manipulation tasks by using a diffusion model to generate task-relevant trajectories. This trajectory-level guidance mitigates error accumulation and improves generalization in out-of-distribution scenarios, outperforming state-of-the-art baselines.
Business Value
Enables robots to learn and perform more complex, multi-step tasks with greater reliability and adaptability, reducing the need for extensive manual programming and data collection.