Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: In this paper, we propose UniGS, a unified map representation and
differentiable framework for high-fidelity multimodal 3D reconstruction based
on 3D Gaussian Splatting. Our framework integrates a CUDA-accelerated
rasterization pipeline capable of rendering photo-realistic RGB images,
geometrically accurate depth maps, consistent surface normals, and semantic
logits simultaneously. We redesign the rasterization to render depth via
differentiable ray-ellipsoid intersection rather than using Gaussian centers,
enabling effective optimization of rotation and scale attribute through
analytic depth gradients. Furthermore, we derive the analytic gradient
formulation for surface normal rendering, ensuring geometric consistency among
reconstructed 3D scenes. To improve computational and storage efficiency, we
introduce a learnable attribute that enables differentiable pruning of
Gaussians with minimal contribution during training. Quantitative and
qualitative experiments demonstrate state-of-the-art reconstruction accuracy
across all modalities, validating the efficacy of our geometry-aware paradigm.
Source code and multimodal viewer will be available on GitHub.
Key Contributions
Introduces UniGS, a unified framework for high-fidelity multimodal 3D reconstruction using 3D Gaussian Splatting. It features a CUDA-accelerated pipeline for rendering RGB, depth, normals, and semantics simultaneously, with novel differentiable rendering techniques for depth and normals, and an efficient pruning mechanism.
Business Value
Enables creation of highly realistic and detailed 3D models from various data sources, crucial for applications in VR/AR, gaming, digital twins, and autonomous systems. The multimodal output simplifies downstream processing.