Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI Researchers,Generative AI Developers,Digital Artists,ML Engineers 2 weeks ago

CADE 2.5 - ZeResFDG: Frequency-Decoupled, Rescaled and Zero-Projected Guidance for SD/SDXL Latent Diffusion Models

computer-vision › diffusion-models
📄 Abstract

Abstract: We introduce CADE 2.5 (Comfy Adaptive Detail Enhancer), a sampler-level guidance stack for SD/SDXL latent diffusion models. The central module, ZeResFDG, unifies (i) frequency-decoupled guidance that reweights low- and high-frequency components of the guidance signal, (ii) energy rescaling that matches the per-sample magnitude of the guided prediction to the positive branch, and (iii) zero-projection that removes the component parallel to the unconditional direction. A lightweight spectral EMA with hysteresis switches between a conservative and a detail-seeking mode as structure crystallizes during sampling. Across SD/SDXL samplers, ZeResFDG improves sharpness, prompt adherence, and artifact control at moderate guidance scales without any retraining. In addition, we employ a training-free inference-time stabilizer, QSilk Micrograin Stabilizer (quantile clamp + depth/edge-gated micro-detail injection), which improves robustness and yields natural high-frequency micro-texture at high resolutions with negligible overhead. For completeness we note that the same rule is compatible with alternative parameterizations (e.g., velocity), which we briefly discuss in the Appendix; however, this paper focuses on SD/SDXL latent diffusion models.
Authors (1)
Denis Rychkovskiy
DZRobo, Independent Researcher
Institutions
🏛️ DZRobo, Independent Researcher
Submitted
October 14, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Introduces ZeResFDG, a sampler-level guidance stack for SD/SDXL models that unifies frequency-decoupled guidance, energy rescaling, and zero-projection. This method improves sharpness, prompt adherence, and artifact control without retraining, and includes a training-free stabilizer for robustness and micro-texture generation.

Business Value

Enhances the quality and controllability of AI-generated images, making diffusion models more practical for professional creative workflows, marketing, and personalized content generation.