Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Visually localizing an image, i.e., estimating its camera pose, requires
building a scene representation that serves as a visual map. The representation
we choose has direct consequences towards the practicability of our system.
Even when starting from mapping images with known camera poses,
state-of-the-art approaches still require hours of mapping time in the worst
case, and several minutes in the best. This work raises the question whether we
can achieve competitive accuracy much faster. We introduce FastForward, a
method that creates a map representation and relocalizes a query image
on-the-fly in a single feed-forward pass. At the core, we represent multiple
mapping images as a collection of features anchored in 3D space. FastForward
utilizes these mapping features to predict image-to-scene correspondences for
the query image, enabling the estimation of its camera pose. We couple
FastForward with image retrieval and achieve state-of-the-art accuracy when
compared to other approaches with minimal map preparation time. Furthermore,
FastForward demonstrates robust generalization to unseen domains, including
challenging large-scale outdoor environments.
Key Contributions
This paper introduces FastForward, a novel feed-forward method for camera localization that achieves competitive accuracy significantly faster than state-of-the-art approaches. It represents a scene as a collection of 3D-anchored image features and enables on-the-fly pose estimation in a single forward pass, drastically reducing mapping and localization times.
Business Value
Enables real-time, accurate camera localization for applications like AR, robotics, and autonomous driving, where speed and precision are critical, potentially reducing hardware costs and improving system responsiveness.