Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Rectified Flows learn ODE vector fields whose trajectories are straight
between source and target distributions, enabling near one-step inference. We
show that this straight-path objective conceals fundamental failure modes:
under deterministic training, low gradient variance drives memorization of
arbitrary training pairings, even when interpolant lines between pairs
intersect. To analyze this mechanism, we study Gaussian-to-Gaussian transport
and use the loss gradient variance across stochastic and deterministic regimes
to characterize which vector fields optimization favors in each setting. We
then show that, in a setting where all interpolating lines intersect, applying
Rectified Flow yields the same specific pairings at inference as during
training. More generally, we prove that a memorizing vector field exists even
when training interpolants intersect, and that optimizing the straight-path
objective converges to this ill-defined field. At inference, deterministic
integration reproduces the exact training pairings. We validate our findings
empirically on the CelebA dataset, confirming that deterministic interpolants
induce memorization, while the injection of small noise restores
generalization.
Authors (4)
Teodora Reu
Sixtine Dromigny
Michael Bronstein
Francisco Vargas
Submitted
October 20, 2025
Key Contributions
This paper reveals that the straight-path objective in Rectified Flows can conceal failure modes driven by low gradient variance, leading to memorization of training pairings. The authors analyze this mechanism using gradient variance in stochastic vs. deterministic regimes and prove the existence of a memorizing vector field that the straight-path objective converges to, explaining why these models can fail unexpectedly.
Business Value
Understanding and mitigating failure modes in generative models can lead to more reliable and robust AI systems for tasks like data synthesis, anomaly detection, and creative content generation.