Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Synthesizing large-scale, explorable, and geometrically accurate 3D urban
scenes is a challenging yet valuable task in providing immersive and embodied
applications. The challenges lie in the lack of large-scale and high-quality
real-world 3D scans for training generalizable generative models. In this
paper, we take an alternative route to create large-scale 3D scenes by
synergizing the readily available satellite imagery that supplies realistic
coarse geometry and the open-domain diffusion model for creating high-quality
close-up appearances. We propose \textbf{Skyfall-GS}, the first city-block
scale 3D scene creation framework without costly 3D annotations, also featuring
real-time, immersive 3D exploration. We tailor a curriculum-driven iterative
refinement strategy to progressively enhance geometric completeness and
photorealistic textures. Extensive experiments demonstrate that Skyfall-GS
provides improved cross-view consistent geometry and more realistic textures
compared to state-of-the-art approaches. Project page:
https://skyfall-gs.jayinnn.dev/
Authors (9)
Jie-Ying Lee
Yi-Ruei Liu
Shr-Ruei Tsai
Wei-Cheng Chang
Chung-Ho Wu
Jiewen Chan
+3 more
Submitted
October 17, 2025
Key Contributions
Skyfall-GS is the first framework to synthesize city-block scale 3D urban scenes without costly 3D annotations by synergizing satellite imagery (for coarse geometry) and diffusion models (for high-quality appearances). It employs a curriculum-driven iterative refinement strategy to enhance geometric completeness and photorealism, enabling real-time, immersive 3D exploration.
Business Value
Enables the creation of detailed, explorable 3D urban environments for applications like virtual city tours, urban planning simulations, and immersive gaming experiences, reducing the need for expensive manual 3D modeling.