TY - GEN
T1 - Wasserstein GAN with quadratic transport cost
AU - Liu, Huidong
AU - Gu, Xianfeng
AU - Samaras, DImitris
N1 - Publisher Copyright: © 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - Wasserstein GANs are increasingly used in Computer Vision applications as they are easier to train. Previous WGAN variants mainly use the l-1 transport cost to compute the Wasserstein distance between the real and synthetic data distributions. The l-1 transport cost restricts the discriminator to be 1-Lipschitz. However, WGANs with l-1 transport cost were recently shown to not always converge. In this paper, we propose WGAN-QC, a WGAN with quadratic transport cost. Based on the quadratic transport cost, we propose an Optimal Transport Regularizer (OTR) to stabilize the training process of WGAN-QC. We prove that the objective of the discriminator during each generator update computes the exact quadratic Wasserstein distance between real and synthetic data distributions. We also prove that WGAN-QC converges to a local equilibrium point with finite discriminator updates per generator update. We show experimentally on a Dirac distribution that WGAN-QC converges, when many of the l-1 cost WGANs fail to [22]. Qualitative and quantitative results on the CelebA, CelebA-HQ, LSUN and the ImageNet dog datasets show that WGAN-QC is better than state-of-art GAN methods. WGAN-QC has much faster runtime than other WGAN variants.
AB - Wasserstein GANs are increasingly used in Computer Vision applications as they are easier to train. Previous WGAN variants mainly use the l-1 transport cost to compute the Wasserstein distance between the real and synthetic data distributions. The l-1 transport cost restricts the discriminator to be 1-Lipschitz. However, WGANs with l-1 transport cost were recently shown to not always converge. In this paper, we propose WGAN-QC, a WGAN with quadratic transport cost. Based on the quadratic transport cost, we propose an Optimal Transport Regularizer (OTR) to stabilize the training process of WGAN-QC. We prove that the objective of the discriminator during each generator update computes the exact quadratic Wasserstein distance between real and synthetic data distributions. We also prove that WGAN-QC converges to a local equilibrium point with finite discriminator updates per generator update. We show experimentally on a Dirac distribution that WGAN-QC converges, when many of the l-1 cost WGANs fail to [22]. Qualitative and quantitative results on the CelebA, CelebA-HQ, LSUN and the ImageNet dog datasets show that WGAN-QC is better than state-of-art GAN methods. WGAN-QC has much faster runtime than other WGAN variants.
UR - https://www.scopus.com/pages/publications/85081907025
U2 - 10.1109/ICCV.2019.00493
DO - 10.1109/ICCV.2019.00493
M3 - Conference contribution
T3 - Proceedings of the IEEE International Conference on Computer Vision
SP - 4831
EP - 4840
BT - Proceedings - 2019 International Conference on Computer Vision, ICCV 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 17th IEEE/CVF International Conference on Computer Vision, ICCV 2019
Y2 - 27 October 2019 through 2 November 2019
ER -