Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Accurate and reliable estimation of biases of low-cost Inertial Measurement
Units (IMU) is a key factor to maintain the resilience of Visual-Inertial
Odometry (VIO), particularly when visual tracking fails in challenging areas.
In such cases, bias estimates from the VIO can deviate significantly from the
real values because of the insufficient or erroneous vision features,
compromising both localization accuracy and system stability. To address this
challenge, we propose a novel plug-and-play module featuring the Inertial Prior
Network (IPNet), which infers an IMU bias prior by implicitly capturing the
motion characteristics of specific platforms. The core idea is inspired
intuitively by the observation that different platforms exhibit distinctive
motion patterns, while the integration of low-cost IMU measurements suffers
from unbounded error that quickly accumulates over time. Therefore, these
specific motion patterns can be exploited to infer the underlying IMU bias. In
this work, we first directly infer the biases prior only using the raw IMU data
using a sliding window approach, eliminating the dependency on recursive bias
estimation combining visual features, thus effectively preventing error
propagation in challenging areas. Moreover, to compensate for the lack of
ground-truth bias in most visual-inertial datasets, we further introduce an
iterative method to compute the mean per-sequence IMU bias for network training
and release it to benefit society. The framework is trained and evaluated
separately on two public datasets and a self-collected dataset. Extensive
experiments show that our method significantly improves localization precision
and robustness.
Authors (7)
Yang Yi
Kunqing Wang
Jinpu Zhang
Zhen Tan
Xiangke Wang
Hui Shen
+1 more
Key Contributions
Proposes a novel plug-and-play module, the Inertial Prior Network (IPNet), that learns to infer IMU bias priors by implicitly capturing platform-specific motion characteristics. This improves the robustness of Visual-Inertial Odometry (VIO), especially when visual tracking fails.
Business Value
Enhanced robustness and accuracy of VIO systems are crucial for reliable navigation in robotics, drones, AR/VR, and autonomous vehicles, especially in GPS-denied or visually challenging environments.