Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Immersive applications call for synthesizing spatiotemporal 4D content from
casual videos without costly 3D supervision. Existing video-to-4D methods
typically rely on manually annotated camera poses, which are labor-intensive
and brittle for in-the-wild footage. Recent warp-then-inpaint approaches
mitigate the need for pose labels by warping input frames along a novel camera
trajectory and using an inpainting model to fill missing regions, thereby
depicting the 4D scene from diverse viewpoints. However, this
trajectory-to-trajectory formulation often entangles camera motion with scene
dynamics and complicates both modeling and inference. We introduce SEE4D, a
pose-free, trajectory-to-camera framework that replaces explicit trajectory
prediction with rendering to a bank of fixed virtual cameras, thereby
separating camera control from scene modeling. A view-conditional video
inpainting model is trained to learn a robust geometry prior by denoising
realistically synthesized warped images and to inpaint occluded or missing
regions across virtual viewpoints, eliminating the need for explicit 3D
annotations. Building on this inpainting core, we design a spatiotemporal
autoregressive inference pipeline that traverses virtual-camera splines and
extends videos with overlapping windows, enabling coherent generation at
bounded per-step complexity. We validate See4D on cross-view video generation
and sparse reconstruction benchmarks. Across quantitative metrics and
qualitative assessments, our method achieves superior generalization and
improved performance relative to pose- or trajectory-conditioned baselines,
advancing practical 4D world modeling from casual videos.
Authors (11)
Dongyue Lu
Ao Liang
Tianxin Huang
Xiao Fu
Yuyang Zhao
Baorui Ma
+5 more
Submitted
October 30, 2025
Key Contributions
SEE4D introduces a pose-free framework for 4D content generation from casual videos, eliminating the need for manual camera pose annotations. It replaces explicit trajectory prediction with rendering to fixed virtual cameras, separating camera control from scene modeling. This approach simplifies the generation process and improves robustness for in-the-wild footage.
Business Value
Enables easier creation of immersive 3D content for VR/AR applications, virtual tours, and telepresence, potentially lowering production costs and barriers to entry. This can enhance user engagement in digital experiences.