Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Understanding and reasoning about complex 3D environments requires structured
scene representations that capture not only objects but also their semantic and
spatial relationships. While recent works on 3D scene graph generation have
leveraged pretrained VLMs without task-specific fine-tuning, they are largely
confined to single-view settings, fail to support incremental updates as new
observations arrive and lack explicit geometric grounding in 3D space, all of
which are essential for embodied scenarios. In this paper, we propose, ZING-3D,
a framework that leverages the vast knowledge of pretrained foundation models
to enable open-vocabulary recognition and generate a rich semantic
representation of the scene in a zero-shot manner while also enabling
incremental updates and geometric grounding in 3D space, making it suitable for
downstream robotics applications. Our approach leverages VLM reasoning to
generate a rich 2D scene graph, which is grounded in 3D using depth
information. Nodes represent open-vocabulary objects with features, 3D
locations, and semantic context, while edges capture spatial and semantic
relations with inter-object distances. Our experiments on scenes from the
Replica and HM3D dataset show that ZING-3D is effective at capturing spatial
and relational knowledge without the need of task-specific training.
Authors (2)
Pranav Saxena
Jimmy Chiun
Submitted
October 24, 2025
Key Contributions
ZING-3D introduces a framework that leverages foundation models for zero-shot, incremental 3D scene graph generation with geometric grounding. This addresses limitations of existing methods by enabling open-vocabulary recognition, handling new observations, and integrating 3D spatial information, crucial for embodied robotics applications.
Business Value
Enables robots and AI systems to better understand and interact with complex 3D environments in real-time, improving navigation, manipulation, and task completion in fields like autonomous driving, logistics, and smart manufacturing.