Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper Robotics Engineers,AI Researchers,Computer Vision Scientists,3D Graphics Developers 1 week ago

ZING-3D: Zero-shot Incremental 3D Scene Graphs via Vision-Language Models

computer-vision › scene-understanding
📄 Abstract

Abstract: Understanding and reasoning about complex 3D environments requires structured scene representations that capture not only objects but also their semantic and spatial relationships. While recent works on 3D scene graph generation have leveraged pretrained VLMs without task-specific fine-tuning, they are largely confined to single-view settings, fail to support incremental updates as new observations arrive and lack explicit geometric grounding in 3D space, all of which are essential for embodied scenarios. In this paper, we propose, ZING-3D, a framework that leverages the vast knowledge of pretrained foundation models to enable open-vocabulary recognition and generate a rich semantic representation of the scene in a zero-shot manner while also enabling incremental updates and geometric grounding in 3D space, making it suitable for downstream robotics applications. Our approach leverages VLM reasoning to generate a rich 2D scene graph, which is grounded in 3D using depth information. Nodes represent open-vocabulary objects with features, 3D locations, and semantic context, while edges capture spatial and semantic relations with inter-object distances. Our experiments on scenes from the Replica and HM3D dataset show that ZING-3D is effective at capturing spatial and relational knowledge without the need of task-specific training.
Authors (2)
Pranav Saxena
Jimmy Chiun
Submitted
October 24, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

ZING-3D introduces a framework that leverages foundation models for zero-shot, incremental 3D scene graph generation with geometric grounding. This addresses limitations of existing methods by enabling open-vocabulary recognition, handling new observations, and integrating 3D spatial information, crucial for embodied robotics applications.

Business Value

Enables robots and AI systems to better understand and interact with complex 3D environments in real-time, improving navigation, manipulation, and task completion in fields like autonomous driving, logistics, and smart manufacturing.