Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Sinkhorn Gradient Flow (2401.14069v1)

Published 25 Jan 2024 in cs.LG

Abstract: Wasserstein Gradient Flows (WGF) with respect to specific functionals have been widely used in the machine learning literature. Recently, neural networks have been adopted to approximate certain intractable parts of the underlying Wasserstein gradient flow and result in efficient inference procedures. In this paper, we introduce the Neural Sinkhorn Gradient Flow (NSGF) model, which parametrizes the time-varying velocity field of the Wasserstein gradient flow w.r.t. the Sinkhorn divergence to the target distribution starting a given source distribution. We utilize the velocity field matching training scheme in NSGF, which only requires samples from the source and target distribution to compute an empirical velocity field approximation. Our theoretical analyses show that as the sample size increases to infinity, the mean-field limit of the empirical approximation converges to the true underlying velocity field. To further enhance model efficiency on high-dimensional tasks, a two-phase NSGF++ model is devised, which first follows the Sinkhorn flow to approach the image manifold quickly ($\le 5$ NFEs) and then refines the samples along a simple straight flow. Numerical experiments with synthetic and real-world benchmark datasets support our theoretical results and demonstrate the effectiveness of the proposed methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (66)
  1. Screening sinkhorn algorithm for regularized optimal transport. NeurIPS, 32, 2019.
  2. Building normalizing flows with stochastic interpolants. In The Eleventh ICLR, 2023.
  3. Optimizing functionals on the space of probabilities with input convex neural networks. Transactions on Machine Learning Research, 2022.
  4. Gradient flows: in metric spaces and in the space of probability measures. Springer Science & Business Media, 2005.
  5. Input convex neural networks. In ICML, pages 146–155. PMLR, 2017.
  6. Refining deep generative models via discriminator gradient flow. In ICLR, 2021.
  7. Maximum mean discrepancy gradient flow. NeurIPS, 32, 2019.
  8. Proximal optimal transport modeling of population dynamics. In International Conference on Artificial Intelligence and Statistics, pages 6511–6528. PMLR, 2022.
  9. John C Butcher. Implicit runge-kutta processes. Mathematics of computation, 18(85):50–64, 1964.
  10. Density ratio estimation via infinitesimal classification. In International Conference on Artificial Intelligence and Statistics, pages 2552–2573. PMLR, 2022.
  11. Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. NeurIPS, 26, 2013.
  12. Sliced iterative normalizing flows. arXiv preprint arXiv:2007.00674, 2020.
  13. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In Proceedings of the European conference on computer vision (ECCV), pages 447–463, 2018.
  14. Image generation with shortest path diffusion. arXiv preprint arXiv:2306.00501, 2023.
  15. End-to-end sinkhorn autoencoder with noise generator. IEEE Access, 9:7211–7219, 2020.
  16. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
  17. Nonparametric generative modeling with conditional sliced-wasserstein flows. 2023.
  18. Variational wasserstein gradient flow. In ICML, pages 6185–6215. PMLR, 2022.
  19. Learning with minibatch wasserstein: asymptotic and gradient properties. arXiv preprint arXiv:1910.04091, 2019.
  20. Unbalanced minibatch optimal transport; applications to domain adaptation. In ICML, pages 3186–3197. PMLR, 2021.
  21. Minibatch optimal transport distances; analysis and applications. arXiv preprint arXiv:2101.01792, 2021.
  22. Interpolating between optimal transport and mmd using sinkhorn divergences. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 2681–2690. PMLR, 2019.
  23. Gerald B Folland. Real analysis: modern techniques and their applications, volume 40. John Wiley & Sons, 1999.
  24. Deep generative learning via variational gradient flow. In ICML, pages 2093–2101. PMLR, 2019.
  25. Deep generative learning via euler particle transport. In Mathematical and Scientific Machine Learning, pages 336–368. PMLR, 2022.
  26. Stochastic optimization for large-scale optimal transport. NeurIPS, 29, 2016.
  27. Learning generative models with sinkhorn divergences. In International Conference on Artificial Intelligence and Statistics, pages 1608–1617. PMLR, 2018.
  28. Generative adversarial nets. NeurIPS, 27, 2014.
  29. Ffjord: Free-form continuous dynamics for scalable reversible generative models. arXiv preprint arXiv:1810.01367, 2018.
  30. Deep generative wasserstein gradient flows. 2022.
  31. Gans trained by a two time-scale update rule converge to a local nash equilibrium. NeurIPS, 30, 2017.
  32. Denoising diffusion probabilistic models. NeurIPS, 33:6840–6851, 2020.
  33. The limits of min-max optimization algorithms: Convergence to spurious non-critical sets. In ICML, pages 4337–4348. PMLR, 2021.
  34. The variational formulation of the fokker–planck equation. SIAM journal on mathematical analysis, 29(1):1–17, 1998.
  35. Iraj Kalantari. Induction over the Continuum, pages 145–154. Springer Netherlands, Dordrecht, 2007.
  36. Multi-source domain adaptation with sinkhorn barycenter. In 2021 29th European Signal Processing Conference (EUSIPCO), pages 1371–1375. IEEE, 2021.
  37. Do neural optimal transport solvers work? a continuous wasserstein-2 benchmark. NeurIPS, 34:14593–14605, 2021.
  38. Mmd gan: Towards deeper understanding of moment matching network. NeurIPS, 30, 2017.
  39. Flow matching for generative modeling. In The Eleventh ICLR, 2023.
  40. Flow straight and fast: Learning to generate and transfer data with rectified flow. In The Eleventh ICLR, 2023.
  41. Sliced-wasserstein flows: Nonparametric generative modeling via optimal transport and diffusions. In ICML, pages 4104–4113. PMLR, 2019.
  42. Sinkhorn barycenters with free support via frank-wolfe algorithm. NeurIPS, 32, 2019.
  43. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602, 2013.
  44. Large-scale wasserstein gradient flows. NeurIPS, 34:15243–15256, 2021.
  45. Unbalanced sobolev descent. NeurIPS, 33:17034–17043, 2020.
  46. Sobolev descent. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 2976–2985. PMLR, 2019.
  47. Fast sinkhorn filters: Using matrix scaling for non-rigid shape correspondence with functional maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 384–393, 2021.
  48. Sinkhorn autoencoders. In Uncertainty in Artificial Intelligence, pages 733–743. PMLR, 2020.
  49. Computational optimal transport. Center for Research in Economics and Statistics Working Papers, (2017-86), 2017.
  50. Numerical solution of stochastic differential equations with jumps in finance, volume 64. Springer Science & Business Media, 2010.
  51. Multisample flow matching: Straightening flows with minibatch couplings. arXiv preprint arXiv:2304.14772, 2023.
  52. Improved techniques for training gans. NeurIPS, 29, 2016.
  53. Filippo Santambrogio. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
  54. Filippo Santambrogio. {{\{{Euclidean, metric, and Wasserstein}}\}} gradient flows: an overview. Bulletin of Mathematical Sciences, 7:87–154, 2017.
  55. On kinetic optimal probability paths for generative models. In ICML, pages 30883–30907. PMLR, 2023.
  56. Sinkhorn barycenter via functional gradient descent. NeurIPS, 33:986–996, 2020.
  57. Mastering the game of go with deep neural networks and tree search. nature, 529(7587):484–489, 2016.
  58. Generative modeling by estimating gradients of the data distribution. NeurIPS, 32, 2019.
  59. Score-based generative modeling through stochastic differential equations. In ICLR, 2021.
  60. An introduction to numerical analysis. Cambridge university press, 2003.
  61. Improving and generalizing flow-based generative models with minibatch optimal transport. In ICML Workshop on New Frontiers in Learning, Control, and Dynamical Systems, 2023.
  62. Cédric Villani et al. Optimal transport: old and new, volume 338. Springer, 2009.
  63. Improving mmd-gan training with repulsive loss function. arXiv preprint arXiv:1812.09916, 2018.
  64. A mean-field games laboratory for generative modeling. arXiv preprint arXiv:2304.13534, 2023.
  65. Dpvi: A dynamic-weight particle-based variational inference framework. arXiv preprint arXiv:2112.00945, 2021.
  66. Wasserstein flow meets replicator dynamics: A mean-field analysis of representation learning in actor-critic. NeurIPS, 34:15993–16006, 2021.
Citations (4)

Summary

We haven't generated a summary for this paper yet.