Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 98% Match Research Paper AI Researchers,ML Engineers,AI Ethicists,Legal Experts in IP 1 month ago

How Diffusion Models Memorize

computer-vision › diffusion-models
📄 Abstract

Abstract: Despite their success in image generation, diffusion models can memorize training data, raising serious privacy and copyright concerns. Although prior work has sought to characterize, detect, and mitigate memorization, the fundamental question of why and how it occurs remains unresolved. In this paper, we revisit the diffusion and denoising process and analyze latent space dynamics to address the question: "How do diffusion models memorize?" We show that memorization is driven by the overestimation of training samples during early denoising, which reduces diversity, collapses denoising trajectories, and accelerates convergence toward the memorized image. Specifically: (i) memorization cannot be explained by overfitting alone, as training loss is larger under memorization due to classifier-free guidance amplifying predictions and inducing overestimation; (ii) memorized prompts inject training images into noise predictions, forcing latent trajectories to converge and steering denoising toward their paired samples; and (iii) a decomposition of intermediate latents reveals how initial randomness is quickly suppressed and replaced by memorized content, with deviations from the theoretical denoising schedule correlating almost perfectly with memorization severity. Together, these results identify early overestimation as the central underlying mechanism of memorization in diffusion models.

Key Contributions

This paper provides a fundamental explanation for how diffusion models memorize training data by analyzing latent space dynamics and the denoising process. It reveals that memorization is driven by the overestimation of training samples during early denoising, which reduces diversity and forces convergence towards memorized images, a phenomenon amplified by classifier-free guidance.

Business Value

Crucial for building trust and ensuring responsible deployment of diffusion models. Understanding memorization helps in developing strategies to protect sensitive training data, comply with copyright, and prevent the generation of harmful or infringing content.