Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper AI researchers,Robotics engineers,AR/VR developers,Computer vision scientists 2 weeks ago

Think with 3D: Geometric Imagination Grounded Spatial Reasoning from Limited Views

computer-vision › 3d-vision
📄 Abstract

Abstract: Though recent advances in vision-language models (VLMs) have achieved remarkable progress across a wide range of multimodal tasks, understanding 3D spatial relationships from limited views remains a significant challenge. Previous reasoning methods typically rely on pure text (e.g., topological cognitive maps) or on 2D visual cues. However, their limited representational capacity hinders performance in specific tasks that require 3D spatial imagination. To address this limitation, we propose 3DThinker, a framework that can effectively exploits the rich geometric information embedded within images while reasoning, like humans do. Our framework is the first to enable 3D mentaling during reasoning without any 3D prior input, and it does not rely on explicitly labeled 3D data for training. Specifically, our training consists of two stages. First, we perform supervised training to align the 3D latent generated by VLM while reasoning with that of a 3D foundation model (e.g., VGGT). Then, we optimize the entire reasoning trajectory solely based on outcome signals, thereby refining the underlying 3D mentaling. Extensive experiments across multiple benchmarks show that 3DThinker consistently outperforms strong baselines and offers a new perspective toward unifying 3D representations into multimodal reasoning. Our code will be available at https://github.com/zhangquanchen/3DThinker.
Authors (10)
Zhangquan Chen
Manyuan Zhang
Xinlei Yu
Xufang Luo
Mingze Sun
Zihao Pan
+4 more
Submitted
October 21, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Proposes 3DThinker, a framework enabling 3D spatial reasoning and '3D mentalizing' from limited views without explicit 3D prior input or labeled 3D data. It effectively leverages geometric information within images for human-like spatial understanding.

Business Value

Crucial for developing more capable robots, AR/VR systems, and autonomous agents that can understand and interact with the physical world in a 3D context.