Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: We introduce CADE 2.5 (Comfy Adaptive Detail Enhancer), a sampler-level
guidance stack for SD/SDXL latent diffusion models. The central module,
ZeResFDG, unifies (i) frequency-decoupled guidance that reweights low- and
high-frequency components of the guidance signal, (ii) energy rescaling that
matches the per-sample magnitude of the guided prediction to the positive
branch, and (iii) zero-projection that removes the component parallel to the
unconditional direction. A lightweight spectral EMA with hysteresis switches
between a conservative and a detail-seeking mode as structure crystallizes
during sampling. Across SD/SDXL samplers, ZeResFDG improves sharpness, prompt
adherence, and artifact control at moderate guidance scales without any
retraining. In addition, we employ a training-free inference-time stabilizer,
QSilk Micrograin Stabilizer (quantile clamp + depth/edge-gated micro-detail
injection), which improves robustness and yields natural high-frequency
micro-texture at high resolutions with negligible overhead. For completeness we
note that the same rule is compatible with alternative parameterizations (e.g.,
velocity), which we briefly discuss in the Appendix; however, this paper
focuses on SD/SDXL latent diffusion models.
Authors (1)
Denis Rychkovskiy
DZRobo, Independent Researcher
Institutions
🏛️ DZRobo, Independent Researcher
Submitted
October 14, 2025
Key Contributions
Introduces ZeResFDG, a sampler-level guidance stack for SD/SDXL models that unifies frequency-decoupled guidance, energy rescaling, and zero-projection. This method improves sharpness, prompt adherence, and artifact control without retraining, and includes a training-free stabilizer for robustness and micro-texture generation.
Business Value
Enhances the quality and controllability of AI-generated images, making diffusion models more practical for professional creative workflows, marketing, and personalized content generation.