Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: 3D Gaussian Splatting SLAM has emerged as a widely used technique for
high-fidelity mapping in spatial intelligence. However, existing methods often
rely on a single representation scheme, which limits their performance in
large-scale dynamic outdoor scenes and leads to cumulative pose errors and
scale ambiguity. To address these challenges, we propose \textbf{LVD-GS}, a
novel LiDAR-Visual 3D Gaussian Splatting SLAM system. Motivated by the human
chain-of-thought process for information seeking, we introduce a hierarchical
collaborative representation module that facilitates mutual reinforcement for
mapping optimization, effectively mitigating scale drift and enhancing
reconstruction robustness. Furthermore, to effectively eliminate the influence
of dynamic objects, we propose a joint dynamic modeling module that generates
fine-grained dynamic masks by fusing open-world segmentation with implicit
residual constraints, guided by uncertainty estimates from DINO-Depth features.
Extensive evaluations on KITTI, nuScenes, and self-collected datasets
demonstrate that our approach achieves state-of-the-art performance compared to
existing methods.
Authors (7)
Wenkai Zhu
Xu Li
Qimin Xu
Benwu Wang
Kun Wei
Yiming Peng
+1 more
Submitted
October 26, 2025
Key Contributions
LVD-GS proposes a LiDAR-Visual 3D Gaussian Splatting SLAM system that addresses challenges in dynamic outdoor scenes. It introduces a hierarchical representation module for mapping optimization and a joint dynamic modeling module to effectively remove dynamic objects, mitigating scale drift and enhancing reconstruction robustness.
Business Value
Enables more accurate and robust 3D mapping for applications like autonomous navigation, AR/VR content creation, and robotic perception in complex, real-world environments.