Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper Computer Graphics Researchers,3D Vision Researchers,Robotics Engineers,VR/AR Developers 3 weeks ago

UniGS: Unified Geometry-Aware Gaussian Splatting for Multimodal Rendering

computer-vision › 3d-vision
📄 Abstract

Abstract: In this paper, we propose UniGS, a unified map representation and differentiable framework for high-fidelity multimodal 3D reconstruction based on 3D Gaussian Splatting. Our framework integrates a CUDA-accelerated rasterization pipeline capable of rendering photo-realistic RGB images, geometrically accurate depth maps, consistent surface normals, and semantic logits simultaneously. We redesign the rasterization to render depth via differentiable ray-ellipsoid intersection rather than using Gaussian centers, enabling effective optimization of rotation and scale attribute through analytic depth gradients. Furthermore, we derive the analytic gradient formulation for surface normal rendering, ensuring geometric consistency among reconstructed 3D scenes. To improve computational and storage efficiency, we introduce a learnable attribute that enables differentiable pruning of Gaussians with minimal contribution during training. Quantitative and qualitative experiments demonstrate state-of-the-art reconstruction accuracy across all modalities, validating the efficacy of our geometry-aware paradigm. Source code and multimodal viewer will be available on GitHub.

Key Contributions

Introduces UniGS, a unified framework for high-fidelity multimodal 3D reconstruction using 3D Gaussian Splatting. It features a CUDA-accelerated pipeline for rendering RGB, depth, normals, and semantics simultaneously, with novel differentiable rendering techniques for depth and normals, and an efficient pruning mechanism.

Business Value

Enables creation of highly realistic and detailed 3D models from various data sources, crucial for applications in VR/AR, gaming, digital twins, and autonomous systems. The multimodal output simplifies downstream processing.