Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper Robotics Researchers,AI Engineers,Embodied AI Researchers,ML Engineers 3 weeks ago

InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy

robotics › embodied-agents
📄 Abstract

Abstract: We introduce InternVLA-M1, a unified framework for spatial grounding and robot control that advances instruction-following robots toward scalable, general-purpose intelligence. Its core idea is spatially guided vision-language-action training, where spatial grounding serves as the critical link between instructions and robot actions. InternVLA-M1 employs a two-stage pipeline: (i) spatial grounding pre-training on over 2.3M spatial reasoning data to determine ``where to act'' by aligning instructions with visual, embodiment-agnostic positions, and (ii) spatially guided action post-training to decide ``how to act'' by generating embodiment-aware actions through plug-and-play spatial prompting. This spatially guided training recipe yields consistent gains: InternVLA-M1 outperforms its variant without spatial guidance by +14.6% on SimplerEnv Google Robot, +17% on WidowX, and +4.3% on LIBERO Franka, while demonstrating stronger spatial reasoning capability in box, point, and trace prediction. To further scale instruction following, we built a simulation engine to collect 244K generalizable pick-and-place episodes, enabling a 6.2% average improvement across 200 tasks and 3K+ objects. In real-world clustered pick-and-place, InternVLA-M1 improved by 7.3%, and with synthetic co-training, achieved +20.6% on unseen objects and novel configurations. Moreover, in long-horizon reasoning-intensive scenarios, it surpassed existing works by over 10%. These results highlight spatially guided training as a unifying principle for scalable and resilient generalist robots. Code and models are available at https://github.com/InternRobotics/InternVLA-M1.
Authors (29)
Xinyi Chen
Yilun Chen
Yanwei Fu
Ning Gao
Jiaya Jia
Weiyang Jin
+23 more
Submitted
October 15, 2025
arXiv Category
cs.RO
arXiv PDF

Key Contributions

Introduces InternVLA-M1, a unified framework for spatial grounding and robot control that enables generalist robot policies. It uses a two-stage pipeline: spatial grounding pre-training to align instructions with visual positions, and spatially guided action post-training to generate embodiment-aware actions, significantly improving instruction-following performance.

Business Value

Enables the development of more versatile and intelligent robots capable of understanding and executing complex tasks based on natural language, accelerating automation in various industries.