Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Imitation learning is a popular method for teaching robots new behaviors.
However, most existing methods focus on teaching short, isolated skills rather
than long, multi-step tasks. To bridge this gap, imitation learning algorithms
must not only learn individual skills but also an abstract understanding of how
to sequence these skills to perform extended tasks effectively. This paper
addresses this challenge by proposing a neuro-symbolic imitation learning
framework. Using task demonstrations, the system first learns a symbolic
representation that abstracts the low-level state-action space. The learned
representation decomposes a task into easier subtasks and allows the system to
leverage symbolic planning to generate abstract plans. Subsequently, the system
utilizes this task decomposition to learn a set of neural skills capable of
refining abstract plans into actionable robot commands. Experimental results in
three simulated robotic environments demonstrate that, compared to baselines,
our neuro-symbolic approach increases data efficiency, improves generalization
capabilities, and facilitates interpretability.
Authors (3)
Leon Keller
Daniel Tanneberg
Jan Peters
Key Contributions
This paper proposes a neuro-symbolic imitation learning framework to address the challenge of teaching robots long, multi-step tasks. The system learns a symbolic representation to decompose tasks into subtasks, enabling symbolic planning for abstract plans. It then uses this decomposition to train neural skills that refine these plans into actionable robot commands. This approach allows robots to learn and execute extended tasks more effectively.
Business Value
Enables robots to learn and perform more complex, sequential tasks autonomously, leading to increased automation capabilities in manufacturing, logistics, and other industries.