Abstract
Generative adversarial networks are popular for generative tasks; however‚ they often require careful architecture selection and extensive empirical tuning‚ and they are prone to mode collapse. To overcome these challenges‚ we propose a novel model that identifies the low-dimensional structure of the underlying data distribution‚ maps it into a low-dimensional latent space while preserving the underlying geometry‚ and then optimally transports a reference measure to the embedded distribution. We prove three key properties of our method: (1) the encoder preserves the geometry of the underlying data; (2) the generator is c-cyclically monotone‚ where c is an intrinsic embedding cost employed by the encoder; and (3) the discriminator's modulus of continuity improves with the geometric preservation of the data. Numerical experiments demonstrate the effectiveness of our approach in generating high-quality images and exhibiting robustness to both mode collapse and training instability.
| Original language | English (US) |
|---|---|
| Pages (from-to) | 1184-1209 |
| Number of pages | 26 |
| Journal | SIAM Journal on Mathematics of Data Science |
| Volume | 7 |
| Issue number | 3 |
| DOIs | |
| State | Published - 2025 |
Bibliographical note
Publisher Copyright:© 2025 Society for Industrial and Applied Mathematics.
Keywords
- GAN‚
- GromovMonge distance‚
- c-cyclical monotonicity‚
- generative adversarial network‚
- geometry-preserving encoder‚
- mal transport‚
- mode collapse‚
- opti-
- training instability