Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper Generative AI Researchers,ML Engineers,Computer Vision Researchers 2 weeks ago

AlignFlow: Improving Flow-based Generative Models with Semi-Discrete Optimal Transport

generative-ai › flow-models
📄 Abstract

Abstract: Flow-based Generative Models (FGMs) effectively transform noise into complex data distributions. Incorporating Optimal Transport (OT) to couple noise and data during FGM training has been shown to improve the straightness of flow trajectories, enabling more effective inference. However, existing OT-based methods estimate the OT plan using (mini-)batches of sampled noise and data points, which limits their scalability to large and high-dimensional datasets in FGMs. This paper introduces AlignFlow, a novel approach that leverages Semi-Discrete Optimal Transport (SDOT) to enhance the training of FGMs by establishing an explicit, optimal alignment between noise distribution and data points with guaranteed convergence. SDOT computes a transport map by partitioning the noise space into Laguerre cells, each mapped to a corresponding data point. During FGM training, i.i.d. noise samples are paired with data points via the SDOT map. AlignFlow scales well to large datasets and model architectures with negligible computational overhead. Experimental results show that AlignFlow improves the performance of a wide range of state-of-the-art FGM algorithms and can be integrated as a plug-and-play component. Code is available at: https://github.com/konglk1203/AlignFlow.
Authors (7)
Lingkai Kong
Molei Tao
Yang Liu
Bryan Wang
Jinmiao Fu
Chien-Chih Wang
+1 more
Submitted
October 16, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

AlignFlow introduces Semi-Discrete Optimal Transport (SDOT) to enhance Flow-based Generative Models (FGMs). By partitioning the noise space into Laguerre cells mapped to data points, SDOT establishes an explicit, optimal alignment between noise and data distributions, offering guaranteed convergence and improved scalability for large, high-dimensional datasets.

Business Value

Enables the generation of higher-quality and more diverse synthetic data, which can be used for training other models, data augmentation, and privacy-preserving data sharing.