Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper ML Researchers,Deep Learning Engineers,Students of Generative Models 2 weeks ago

Gradient Variance Reveals Failure Modes in Flow-Based Generative Models

generative-ai › flow-models
📄 Abstract

Abstract: Rectified Flows learn ODE vector fields whose trajectories are straight between source and target distributions, enabling near one-step inference. We show that this straight-path objective conceals fundamental failure modes: under deterministic training, low gradient variance drives memorization of arbitrary training pairings, even when interpolant lines between pairs intersect. To analyze this mechanism, we study Gaussian-to-Gaussian transport and use the loss gradient variance across stochastic and deterministic regimes to characterize which vector fields optimization favors in each setting. We then show that, in a setting where all interpolating lines intersect, applying Rectified Flow yields the same specific pairings at inference as during training. More generally, we prove that a memorizing vector field exists even when training interpolants intersect, and that optimizing the straight-path objective converges to this ill-defined field. At inference, deterministic integration reproduces the exact training pairings. We validate our findings empirically on the CelebA dataset, confirming that deterministic interpolants induce memorization, while the injection of small noise restores generalization.
Authors (4)
Teodora Reu
Sixtine Dromigny
Michael Bronstein
Francisco Vargas
Submitted
October 20, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper reveals that the straight-path objective in Rectified Flows can conceal failure modes driven by low gradient variance, leading to memorization of training pairings. The authors analyze this mechanism using gradient variance in stochastic vs. deterministic regimes and prove the existence of a memorizing vector field that the straight-path objective converges to, explaining why these models can fail unexpectedly.

Business Value

Understanding and mitigating failure modes in generative models can lead to more reliable and robust AI systems for tasks like data synthesis, anomaly detection, and creative content generation.