Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper Researchers in 3D vision and robotics,Developers of autonomous systems,Engineers working with event cameras 3 weeks ago

DEGS: Deformable Event-based 3D Gaussian Splatting from RGB and Event Stream

computer-vision › 3d-vision
📄 Abstract

Abstract: Reconstructing Dynamic 3D Gaussian Splatting (3DGS) from low-framerate RGB videos is challenging. This is because large inter-frame motions will increase the uncertainty of the solution space. For example, one pixel in the first frame might have more choices to reach the corresponding pixel in the second frame. Event cameras can asynchronously capture rapid visual changes and are robust to motion blur, but they do not provide color information. Intuitively, the event stream can provide deterministic constraints for the inter-frame large motion by the event trajectories. Hence, combining low-temporal-resolution images with high-framerate event streams can address this challenge. However, it is challenging to jointly optimize Dynamic 3DGS using both RGB and event modalities due to the significant discrepancy between these two data modalities. This paper introduces a novel framework that jointly optimizes dynamic 3DGS from the two modalities. The key idea is to adopt event motion priors to guide the optimization of the deformation fields. First, we extract the motion priors encoded in event streams by using the proposed LoCM unsupervised fine-tuning framework to adapt an event flow estimator to a certain unseen scene. Then, we present the geometry-aware data association method to build the event-Gaussian motion correspondence, which is the primary foundation of the pipeline, accompanied by two useful strategies, namely motion decomposition and inter-frame pseudo-label. Extensive experiments show that our method outperforms existing image and event-based approaches across synthetic and real scenes and prove that our method can effectively optimize dynamic 3DGS with the help of event data.

Key Contributions

DEGS introduces a novel framework for jointly optimizing dynamic 3D Gaussian Splatting from both RGB and event camera streams. It addresses the challenge of large inter-frame motions in RGB videos by leveraging the precise temporal constraints provided by event camera data, enabling more robust reconstruction of dynamic scenes.

Business Value

Improves the ability of systems (e.g., robots, autonomous vehicles) to perceive and reconstruct dynamic environments in real-time, enhancing safety and operational capabilities.