Monotone Generative Modeling via a Gromov-Monge Embedding

Research output: Contribution to journalArticlepeer-review

Abstract

Generative adversarial networks are popular for generative tasks; however‚ they often require careful architecture selection and extensive empirical tuning‚ and they are prone to mode collapse. To overcome these challenges‚ we propose a novel model that identifies the low-dimensional structure of the underlying data distribution‚ maps it into a low-dimensional latent space while preserving the underlying geometry‚ and then optimally transports a reference measure to the embedded distribution. We prove three key properties of our method: (1) the encoder preserves the geometry of the underlying data; (2) the generator is c-cyclically monotone‚ where c is an intrinsic embedding cost employed by the encoder; and (3) the discriminator's modulus of continuity improves with the geometric preservation of the data. Numerical experiments demonstrate the effectiveness of our approach in generating high-quality images and exhibiting robustness to both mode collapse and training instability.

Original languageEnglish (US)
Pages (from-to)1184-1209
Number of pages26
JournalSIAM Journal on Mathematics of Data Science
Volume7
Issue number3
DOIs
StatePublished - 2025

Bibliographical note

Publisher Copyright:
© 2025 Society for Industrial and Applied Mathematics.

Keywords

  • GAN‚
  • GromovMonge distance‚
  • c-cyclical monotonicity‚
  • generative adversarial network‚
  • geometry-preserving encoder‚
  • mal transport‚
  • mode collapse‚
  • opti-
  • training instability

Fingerprint

Dive into the research topics of 'Monotone Generative Modeling via a Gromov-Monge Embedding'. Together they form a unique fingerprint.

Cite this