Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Humanoid robots are promising to acquire various skills by imitating human
behaviors. However, existing algorithms are only capable of tracking smooth,
low-speed human motions, even with delicate reward and curriculum design. This
paper presents a physics-based humanoid control framework, aiming to master
highly-dynamic human behaviors such as Kungfu and dancing through multi-steps
motion processing and adaptive motion tracking. For motion processing, we
design a pipeline to extract, filter out, correct, and retarget motions, while
ensuring compliance with physical constraints to the maximum extent. For motion
imitation, we formulate a bi-level optimization problem to dynamically adjust
the tracking accuracy tolerance based on the current tracking error, creating
an adaptive curriculum mechanism. We further construct an asymmetric
actor-critic framework for policy training. In experiments, we train whole-body
control policies to imitate a set of highly-dynamic motions. Our method
achieves significantly lower tracking errors than existing approaches and is
successfully deployed on the Unitree G1 robot, demonstrating stable and
expressive behaviors. The project page is https://kungfu-bot.github.io.
Authors (9)
Weiji Xie
Jinrui Han
Jiakun Zheng
Huanyu Li
Xinzhe Liu
Jiyuan Shi
+3 more
Key Contributions
Presents a physics-based humanoid control framework for learning highly-dynamic skills like Kungfu and dancing. It features a multi-step motion processing pipeline and an adaptive motion tracking mechanism using bi-level optimization, enabling humanoid robots to imitate complex, fast-paced human behaviors.
Business Value
Enables the development of more versatile and capable humanoid robots for applications requiring agility and complex movements, such as entertainment, assistance, or hazardous environment operations.