Help Privacy Policy Disclaimer
  Advanced SearchBrowse





Primal-Dual Wasserstein GAN


Akata,  Zeynep
Computer Vision and Multimodal Computing, MPI for Informatics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (public)

(Preprint), 3MB

Supplementary Material (public)
There is no public supplementary material available

Gemici, M., Akata, Z., & Welling, M. (2018). Primal-Dual Wasserstein GAN. Retrieved from http://arxiv.org/abs/1805.09575.

Cite as: http://hdl.handle.net/21.11116/0000-0002-18D3-5
We introduce Primal-Dual Wasserstein GAN, a new learning algorithm for building latent variable models of the data distribution based on the primal and the dual formulations of the optimal transport (OT) problem. We utilize the primal formulation to learn a flexible inference mechanism and to create an optimal approximate coupling between the data distribution and the generative model. In order to learn the generative model, we use the dual formulation and train the decoder adversarially through a critic network that is regularized by the approximate coupling obtained from the primal. Unlike previous methods that violate various properties of the optimal critic, we regularize the norm and the direction of the gradients of the critic function. Our model shares many of the desirable properties of auto-encoding models in terms of mode coverage and latent structure, while avoiding their undesirable averaging properties, e.g. their inability to capture sharp visual features when modeling real images. We compare our algorithm with several other generative modeling techniques that utilize Wasserstein distances on Frechet Inception Distance (FID) and Inception Scores (IS).