Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Data scarcity remains a fundamental bottleneck for embodied intelligence.
Existing approaches use large language models (LLMs) to automate gripper-based
simulation generation, but they transfer poorly to dexterous manipulation,
which demands more specialized environment design. Meanwhile, dexterous
manipulation tasks are inherently more difficult due to their higher degrees of
freedom. Massively generating feasible and trainable dexterous hand tasks
remains an open challenge. To this end, we present GenDexHand, a generative
simulation pipeline that autonomously produces diverse robotic tasks and
environments for dexterous manipulation. GenDexHand introduces a closed-loop
refinement process that adjusts object placements and scales based on
vision-language model (VLM) feedback, substantially improving the average
quality of generated environments. Each task is further decomposed into
sub-tasks to enable sequential reinforcement learning, reducing training time
and increasing success rates. Our work provides a viable path toward scalable
training of diverse dexterous hand behaviors in embodied intelligence by
offering a simulation-based solution to synthetic data generation. Our website:
https://winniechen2002.github.io/GenDexHand/.
Key Contributions
GenDexHand is a generative simulation pipeline that autonomously creates diverse robotic tasks and environments for dexterous manipulation. It uses a closed-loop refinement process with VLM feedback to improve environment quality and decomposes tasks for sequential RL, addressing the data scarcity bottleneck in embodied intelligence.
Business Value
Accelerates the development and training of robotic systems, especially for complex manipulation tasks, by automating the creation of realistic and diverse simulation environments. This reduces the need for expensive real-world data collection and manual environment design.