Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper 3D Artists,Game Developers,VR/AR Developers,Robotics Researchers,AI Engineers 2 weeks ago

Direct Numerical Layout Generation for 3D Indoor Scene Synthesis via Spatial Reasoning

computer-vision › 3d-vision
📄 Abstract

Abstract: Realistic 3D indoor scene synthesis is vital for embodied AI and digital content creation. It can be naturally divided into two subtasks: object generation and layout generation. While recent generative models have significantly advanced object-level quality and controllability, layout generation remains challenging due to limited datasets. Existing methods either overfit to these datasets or rely on predefined constraints to optimize numerical layout that sacrifice flexibility. As a result, they fail to generate scenes that are both open-vocabulary and aligned with fine-grained user instructions. We introduce DirectLayout, a framework that directly generates numerical 3D layouts from text descriptions using generalizable spatial reasoning of large language models (LLMs). DirectLayout decomposes the generation into three stages: producing a Bird's-Eye View (BEV) layout, lifting it into 3D space, and refining object placements. To enable explicit spatial reasoning and help the model grasp basic principles of object placement, we employ Chain-of-Thought (CoT) Activation based on the 3D-Front dataset. Additionally, we design CoT-Grounded Generative Layout Reward to enhance generalization and spatial planning. During inference, DirectLayout addresses asset-layout mismatches via Iterative Asset-Layout Alignment through in-context learning. Extensive experiments demonstrate that DirectLayout achieves impressive semantic consistency, generalization and physical plausibility.
Authors (5)
Xingjian Ran
Yixuan Li
Linning Xu
Mulin Yu
Bo Dai
Submitted
June 5, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Introduces DirectLayout, a framework that directly generates numerical 3D indoor scene layouts from text descriptions using LLM spatial reasoning. It decomposes generation into BEV layout, 3D lifting, and refinement stages, enabling open-vocabulary and instruction-aligned scene synthesis, overcoming limitations of dataset overfitting and rigid constraints in prior methods.

Business Value

Accelerates the creation of virtual environments for gaming, VR/AR experiences, and architectural visualization, reducing manual effort and enabling more dynamic content generation.