Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper Computer vision researchers,Robotics engineers,AR/VR developers,3D artists 2 weeks ago

FLARE: Feed-forward Geometry, Appearance and Camera Estimation from Uncalibrated Sparse Views

computer-vision › 3d-vision
📄 Abstract

Abstract: We present FLARE, a feed-forward model designed to infer high-quality camera poses and 3D geometry from uncalibrated sparse-view images (i.e., as few as 2-8 inputs), which is a challenging yet practical setting in real-world applications. Our solution features a cascaded learning paradigm with camera pose serving as the critical bridge, recognizing its essential role in mapping 3D structures onto 2D image planes. Concretely, FLARE starts with camera pose estimation, whose results condition the subsequent learning of geometric structure and appearance, optimized through the objectives of geometry reconstruction and novel-view synthesis. Utilizing large-scale public datasets for training, our method delivers state-of-the-art performance in the tasks of pose estimation, geometry reconstruction, and novel view synthesis, while maintaining the inference efficiency (i.e., less than 0.5 seconds). The project page and code can be found at: https://zhanghe3z.github.io/FLARE/
Authors (8)
Shangzhan Zhang
Jianyuan Wang
Yinghao Xu
Nan Xue
Christian Rupprecht
Xiaowei Zhou
+2 more
Submitted
February 17, 2025
arXiv Category
cs.CV
arXiv PDF Code

Key Contributions

FLARE is a feed-forward model that infers high-quality camera poses and 3D geometry from uncalibrated sparse views (2-8 images). It employs a cascaded learning paradigm where camera pose estimation guides subsequent geometry and appearance learning, achieving state-of-the-art performance in pose estimation, reconstruction, and novel view synthesis with efficient inference.

Business Value

Enables more accessible and efficient 3D scene understanding and reconstruction, crucial for AR/VR applications, robotics, and digital content creation.

View Code on GitHub