Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
ScaleDiff is a model-agnostic and highly efficient framework for extending the resolution of pretrained diffusion models without additional training. It introduces Neighborhood Patch Attention (NPA) to reduce computational redundancy and Latent Frequency Mixing (LFM) for better detail generation, achieving state-of-the-art performance among training-free methods.
Enables the creation of high-quality, high-resolution images from text prompts more efficiently. This is valuable for industries like advertising, gaming, film, and design, where visual content is paramount.