Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Diffusion models have demonstrated promising performance in real-world video
super-resolution (VSR). However, the dozens of sampling steps they require,
make inference extremely slow. Sampling acceleration techniques, particularly
single-step, provide a potential solution. Nonetheless, achieving one step in
VSR remains challenging, due to the high training overhead on video data and
stringent fidelity demands. To tackle the above issues, we propose DOVE, an
efficient one-step diffusion model for real-world VSR. DOVE is obtained by
fine-tuning a pretrained video diffusion model (i.e., CogVideoX). To
effectively train DOVE, we introduce the latent-pixel training strategy. The
strategy employs a two-stage scheme to gradually adapt the model to the video
super-resolution task. Meanwhile, we design a video processing pipeline to
construct a high-quality dataset tailored for VSR, termed HQ-VSR. Fine-tuning
on this dataset further enhances the restoration capability of DOVE. Extensive
experiments show that DOVE exhibits comparable or superior performance to
multi-step diffusion-based VSR methods. It also offers outstanding inference
efficiency, achieving up to a 28$\times$ speed-up over existing methods such as
MGLD-VSR. Code is available at: https://github.com/zhengchen1999/DOVE.
Authors (7)
Zheng Chen
Zichen Zou
Kewei Zhang
Xiongfei Su
Xin Yuan
Yong Guo
+1 more
Key Contributions
Introduces DOVE, an efficient one-step diffusion model for real-world Video Super-Resolution (VSR) that significantly accelerates inference. It addresses the challenges of high training overhead and stringent fidelity demands by employing a latent-pixel training strategy and constructing a tailored dataset (HQ-VSR).
Business Value
Enables real-time or near-real-time enhancement of low-quality videos, improving viewer experience for streaming services, archival footage restoration, and content creation pipelines.