Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unbalancedness in Neural Monge Maps Improves Unpaired Domain Translation (2311.15100v2)

Published 25 Nov 2023 in cs.CV, cs.AI, and cs.LG

Abstract: In optimal transport (OT), a Monge map is known as a mapping that transports a source distribution to a target distribution in the most cost-efficient way. Recently, multiple neural estimators for Monge maps have been developed and applied in diverse unpaired domain translation tasks, e.g. in single-cell biology and computer vision. However, the classic OT framework enforces mass conservation, which makes it prone to outliers and limits its applicability in real-world scenarios. The latter can be particularly harmful in OT domain translation tasks, where the relative position of a sample within a distribution is explicitly taken into account. While unbalanced OT tackles this challenge in the discrete setting, its integration into neural Monge map estimators has received limited attention. We propose a theoretically grounded method to incorporate unbalancedness into any Monge map estimator. We improve existing estimators to model cell trajectories over time and to predict cellular responses to perturbations. Moreover, our approach seamlessly integrates with the OT flow matching (OT-FM) framework. While we show that OT-FM performs competitively in image translation, we further improve performance by incorporating unbalancedness (UOT-FM), which better preserves relevant features. We hence establish UOT-FM as a principled method for unpaired image translation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (74)
  1. Building normalizing flows with stochastic interpolants, 2023.
  2. Brandon Amos. On amortizing convex conjugates for optimal transport. arXiv preprint arXiv:2210.12153, 2022.
  3. Input Convex Neural Networks. In International Conference on Machine Learning (ICML), volume 34, 2017.
  4. Meta optimal transport, 2023.
  5. The DeepMind JAX Ecosystem, 2020. URL http://github.com/deepmind.
  6. Comprehensive single cell mrna profiling reveals a detailed roadmap for pancreatic endocrinogenesis. Development, 146(12):dev173849, 2019.
  7. Jean-David Benamou. Numerical resolution of an “unbalanced” mass transport problem. ESAIM: Mathematical Modelling and Numerical Analysis, 37(05):851–868, 2003.
  8. Generalizing rna velocity to transient cell states through dynamical modeling. bioRxiv, 2019. doi: 10.1101/820936. URL https://www.biorxiv.org/content/early/2019/10/29/820936.
  9. Diffusion schrödinger bridge with applications to score-based generative modeling, 2023.
  10. JAX: composable transformations of Python+NumPy programs, 2018. URL http://github.com/google/jax.
  11. Yann Brenier. Décomposition polaire et réarrangement monotone des champs de vecteurs. CR Acad. Sci. Paris Sér. I Math., 305, 1987.
  12. Learning Single-Cell Perturbation Responses using Neural Optimal Transport. bioRxiv, 2021.
  13. Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018.
  14. Unbalanced optimal transport: geometry and Kantorovich formulation. Journal of Functional Analysis, 274(11):3090–3123, 2018a.
  15. Scaling algorithms for unbalanced optimal transport problems. Mathematics of Computation, 87(314):2563–2609, 2018b.
  16. Generative modeling through the semi-dual formulation of unbalanced optimal transport, 2023.
  17. Ilvr: Conditioning method for denoising diffusion probabilistic models, 2021.
  18. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation, 2018.
  19. Emnist: an extension of mnist to handwritten letters, 2017.
  20. Marco Cuturi. Sinkhorn Distances: Lightspeed Computation of Optimal Transport. In Advances in Neural Information Processing Systems (NeurIPS), volume 26, 2013.
  21. Optimal Transport Tools (OTT): A JAX Toolbox for all things Wasserstein. arXiv Preprint arXiv:2201.12324, 2022.
  22. Flow matching in latent space, 2023.
  23. Diffusion models beat gans on image synthesis, 2021.
  24. An image is worth 16x16 words: Transformers for image recognition at scale, 2021.
  25. Unbalanced minibatch optimal transport; applications to domain adaptation, 2021.
  26. Interpolating between Optimal Transport and MMD using Sinkhorn Divergences. In International Conference on Artificial Intelligence and Statistics (AISTATS), volume 22, 2019.
  27. Partial neural optimal transport. arXiv preprint arXiv:2303.07988, 2023.
  28. Multiplexed protein maps link subcellular organization to cellular state. Science (New York, N.Y.), 361, 08 2018. doi: 10.1126/science.aar7042.
  29. Gans trained by a two time-scale update rule converge to a local nash equilibrium, 2018.
  30. Machine learning for perturbational single-cell omics. Cell Systems, 12(6):522–537, 2021.
  31. L Kantorovich. On the transfer of masses (in russian). In Doklady Akademii Nauk, volume 37, pp.  227, 1942.
  32. Patrick Kidger. On Neural Differential Equations. PhD thesis, University of Oxford, 2021.
  33. Adam: A Method for Stochastic Optimization. In International Conference on Learning Representations (ICLR), 2014.
  34. Mapping cells through time and space with moscot. bioRxiv, pp.  2023–05, 2023.
  35. Wasserstein-2 generative networks. In International Conference on Learning Representations, 2020.
  36. Do neural optimal transport solvers work? a continuous wasserstein-2 benchmark. 2021. doi: 10.48550/ARXIV.2106.01954. URL https://arxiv.org/abs/2106.01954.
  37. Neural optimal transport. 2022. doi: 10.48550/ARXIV.2201.12220. URL https://arxiv.org/abs/2201.12220.
  38. Cifar-10 (canadian institute for advanced research). URL http://www.cs.toronto.edu/~kriz/cifar.html.
  39. CellRank for directed single-cell fate mapping. Nat. Methods, 19(2):159–170, February 2022.
  40. Parallel unbalanced optimal transport regularization for large scale imaging problems, 2020.
  41. Optimal entropy-transport problems and a new hellinger–kantorovich distance between positive measures. Inventiones Mathematicae, 211(3):969–1117, 2018.
  42. Flow matching for generative modeling, 2023.
  43. I22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPTsb: Image-to-image schrödinger bridge, 2023.
  44. Flow straight and fast: Learning to generate and transfer data with rectified flow, 2022.
  45. Deep learning face attributes in the wild, 2015.
  46. Neural unbalanced optimal transport via cycle-consistent semi-couplings. arXiv preprint arXiv:2209.15621, 2022.
  47. Optimal transport mapping via input convex neural networks. In International Conference on Machine Learning (ICML), volume 37, 2020.
  48. Sdedit: Guided image synthesis and editing with stochastic differential equations, 2022.
  49. Energy-guided entropic neural optimal transport. arXiv preprint arXiv:2304.06094, 2023.
  50. Gaspard Monge. Mémoire sur la théorie des déblais et des remblais. Histoire de l’Académie Royale des Sciences, pp.  666–704, 1781.
  51. Breaking the cycle – colleagues are all you need, 2020.
  52. Contrastive learning for unpaired image-to-image translation, 2020.
  53. Computational Optimal Transport. Foundations and Trends in Machine Learning, 11(5-6), 2019. ISSN 1935-8245.
  54. Unbalanced optimal transport: A unified framework for object detection, 2023.
  55. Multisample flow matching: Straightening flows with minibatch couplings. arXiv preprint arXiv:2304.14772, 2023a.
  56. Neural optimal transport with lagrangian costs. In ICML Workshop on New Frontiers in Learning, Control, and Dynamical Systems, 2023b. URL https://openreview.net/forum?id=myb0FKB8C9.
  57. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022. URL https://github.com/CompVis/latent-diffusionhttps://arxiv.org/abs/2112.10752.
  58. Filippo Santambrogio. Optimal Transport for Applied Mathematicians. Birkhäuser, NY, 55(58-63):94, 2015.
  59. Unit-ddpm: Unpaired image translation with denoising diffusion probabilistic models, 2021.
  60. Optimal-Transport Analysis of Single-Cell Gene Expression Identifies Developmental Trajectories in Reprogramming. Cell, 176(4), 2019.
  61. Unbalanced optimal transport, from theory to numerics. arXiv preprint arXiv:2211.08775, 2022.
  62. Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020.
  63. Massively multiplex chemical transcriptomics at single-cell resolution. Science, 367(6473), 2020.
  64. Unbalanced optimal transport, from theory to numerics, 2023.
  65. Mlp-mixer: An all-mlp architecture for vision, 2021.
  66. TrajectoryNet: A Dynamic Optimal Transport Network for Modeling Cellular Dynamics. In International Conference on Machine Learning (ICML), 2020.
  67. Simulation-free schr\\\backslash\” odinger bridges via score and flow matching. arXiv preprint arXiv:2307.03672, 2023a.
  68. Improving and generalizing flow-based generative models with minibatch optimal transport, 2023b.
  69. Uvcgan: Unet vision transformer cycle-consistent gan for unpaired image-to-image translation, 2022.
  70. The monge gap: A regularizer to learn all transport maps, 2023.
  71. Scalable unbalanced optimal transport using generative adversarial networks, 2019.
  72. Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations, 2022.
  73. Unpaired image-to-image translation using adversarial consistency loss, 2021.
  74. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Computer Vision (ICCV), 2017 IEEE International Conference on, 2017.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Luca Eyring (3 papers)
  2. Dominik Klein (9 papers)
  3. Théo Uscidda (6 papers)
  4. Giovanni Palla (3 papers)
  5. Niki Kilbertus (41 papers)
  6. Zeynep Akata (144 papers)
  7. Fabian Theis (16 papers)
Citations (11)

Summary

  • The paper introduces unbalanced optimal transport into neural Monge maps to overcome mass conservation limits in unpaired domain translation.
  • It demonstrates significant improvements by accurately aligning distributions in tasks like single-cell trajectory inference, drug perturbation modeling, and image translation.
  • Empirical results reveal higher transition correctness and competitive FID scores, highlighting the method's potential for diverse applications.

Analysis of "Unbalancedness in Neural Monge Maps Improves Unpaired Domain Translation"

The paper under examination, "Unbalancedness in Neural Monge Maps Improves Unpaired Domain Translation," explores the integration of unbalanced optimal transport (UOT) into neural Monge map estimators to enhance the performance in unpaired domain translation tasks. The authors argue that traditional optimal transport (OT) assumptions, which require mass conservation, limit the applicability of neural Monge maps in real-world scenarios due to their sensitivity to outliers and distribution shifts. By incorporating unbalancedness, they propose a method that significantly reduces these limitations.

The paper begins with an overview of OT, emphasizing the Monge map as an efficient means of transporting a source distribution to a target distribution. While neural Monge map estimators have been applied successfully in tasks such as computer vision and single-cell biology, the authors point out that traditional OT frameworks enforce mass conservation, complicating applications that involve outliers or imbalances. Their solution is to introduce unbalanced OT, which removes this constraint by allowing for penalization of mass deviations.

With this new methodology, the authors propose a theoretically justified scheme that can incorporate unbalancedness into any existing Monge map estimator. Their framework involves re-scaling the source and target distributions, allowing for a more adaptable and potentially more accurate translation of data between distributions. The authors validate their approach in several key domains.

Key Numerical Results and Empirical Evaluations

  1. Single-Cell Data Processing: The application of UOT to single-cell trajectory inference demonstrates significant improvements over traditional OT. The authors provide quantitative comparisons showing that unbalanced Monge maps yield a higher percentage of correct cell type transitions. Enhanced performance was particularly notable in predicting transitions for the Ngn3 EP cell population, which traditionally suffered from distribution shifts.
  2. Cellular Responses to Drug Perturbations: In the context of modeling drug perturbations, UOT again shows superiority. The empirical results indicate a robust enhancement in predictive performance across multiple drugs, as measured by the Sinkhorn divergence and L2 drug signature.
  3. Unpaired Image Translation: When applied to image translation tasks, especially with the use of the EMNIST dataset, UOT shows a marked improvement in aligning digits with corresponding letters, despite the inherent distribution shifts. For complex tasks on the CelebA dataset, UOT-FM leverages the unbalanced OT framework to achieve competitive FID scores and supports the preservation of relevant input features.

Implications and Future Directions

The integration of unbalanced OT in neural Monge maps opens up new avenues in domain translation applications where traditional methodological limitations exist due to distribution shifts or outliers. This paper's approach could have a broad impact on various fields where OT-based methods are employed. Furthermore, the use of UOT in mini-batch training scenarios could lead to more stable and quicker convergence of neural models, a promising insight for deep learning applications.

Future research could explore the efficacy of different cost functions tailored specifically to various data domains and geometries, as the paper primarily employs the squared Euclidean distance due to its tractability. Moreover, while the empirical results are promising, further work could focus on refining hyperparameter selection, especially concerning unbalancedness parameters, which currently require grid search optimization.

In conclusion, the paper's contribution to the field of optimal transport and neural estimation is significant, providing both theoretical insights and practical improvements across multiple challenging tasks. The proposed framework not only enhances existing methods but also provides a strong foundation for developing even more advanced and adaptable domain translation models in the future.

X Twitter Logo Streamline Icon: https://streamlinehq.com