📄 Abstract
Abstract: Images captured in challenging environments--such as nighttime, smoke, rainy
weather, and underwater--often suffer from significant degradation, resulting
in a substantial loss of visual quality. The effective restoration of these
degraded images is critical for the subsequent vision tasks. While many
existing approaches have successfully incorporated specific priors for
individual tasks, these tailored solutions limit their applicability to other
degradations. In this work, we propose a universal network architecture, dubbed
``ReviveDiff'', which can address various degradations and bring images back to
life by enhancing and restoring their quality. Our approach is inspired by the
observation that, unlike degradation caused by movement or electronic issues,
quality degradation under adverse conditions primarily stems from natural media
(such as fog, water, and low luminance), which generally preserves the original
structures of objects. To restore the quality of such images, we leveraged the
latest advancements in diffusion models and developed ReviveDiff to restore
image quality from both macro and micro levels across some key factors
determining image quality, such as sharpness, distortion, noise level, dynamic
range, and color accuracy. We rigorously evaluated ReviveDiff on seven
benchmark datasets covering five types of degrading conditions: Rainy,
Underwater, Low-light, Smoke, and Nighttime Hazy. Our experimental results
demonstrate that ReviveDiff outperforms the state-of-the-art methods both
quantitatively and visually.
Authors (5)
Wenfeng Huang
Guoan Xu
Wenjing Jia
Stuart Perry
Guangwei Gao
Submitted
September 27, 2024
Key Contributions
ReviveDiff proposes a universal network architecture based on diffusion models for restoring images degraded by various adverse conditions (night, smoke, rain, underwater). It leverages the observation that such degradations preserve original structures, enabling effective quality enhancement.
Business Value
Enables the recovery of usable visual information from challenging environments, improving the performance of downstream tasks like object detection and recognition, and enhancing visual quality for human perception.