Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Imitation learning has proven effective for training robots to perform
complex tasks from expert human demonstrations. However, it remains limited by
its reliance on high-quality, task-specific data, restricting adaptability to
the diverse range of real-world object configurations and scenarios. In
contrast, non-expert data -- such as play data, suboptimal demonstrations,
partial task completions, or rollouts from suboptimal policies -- can offer
broader coverage and lower collection costs. However, conventional imitation
learning approaches fail to utilize this data effectively. To address these
challenges, we posit that with right design decisions, offline reinforcement
learning can be used as a tool to harness non-expert data to enhance the
performance of imitation learning policies. We show that while standard offline
RL approaches can be ineffective at actually leveraging non-expert data under
the sparse data coverage settings typically encountered in the real world,
simple algorithmic modifications can allow for the utilization of this data,
without significant additional assumptions. Our approach shows that broadening
the support of the policy distribution can allow imitation algorithms augmented
by offline RL to solve tasks robustly, showing considerably enhanced recovery
and generalization behavior. In manipulation tasks, these innovations
significantly increase the range of initial conditions where learned policies
are successful when non-expert data is incorporated. Moreover, we show that
these methods are able to leverage all collected data, including partial or
suboptimal demonstrations, to bolster task-directed policy performance. This
underscores the importance of algorithmic techniques for using non-expert data
for robust policy learning in robotics.
Authors (12)
Kevin Huang
Rosario Scalise
Cleah Winston
Ayush Agrawal
Yunchu Zhang
Rohan Baijal
+6 more
Submitted
October 22, 2025
Key Contributions
This paper proposes using offline reinforcement learning to effectively harness non-expert data (play data, suboptimal demonstrations) to enhance imitation learning policies. It addresses the limitations of traditional imitation learning, which relies on high-quality expert data, by showing how offline RL can leverage broader, lower-cost data for improved robustness and adaptability in robotics.
Business Value
Enables more cost-effective and robust training of robotic systems by utilizing readily available, diverse data sources, leading to faster deployment and wider applicability.