Machine Learning

On the regularization of Wasserstein GANs

This topic contains 0 replies, has 1 voice, and was last updated by  arXiv 1 year, 3 months ago.

  • arXiv
    5 pts

    On the regularization of Wasserstein GANs

    Since their invention, generative adversarial networks (GANs) have become a popular approach for learning to model a distribution of real (unlabeled) data. Convergence problems during training are overcome by Wasserstein GANs which minimize the distance between the model and the empirical distribution in terms of a different metric, but thereby introduce a Lipschitz constraint into the optimization problem. A simple way to enforce the Lipschitz constraint on the class of functions, which can be modeled by the neural network, is weight clipping. It was proposed that training can be improved by instead augmenting the loss by a regularization term that penalizes the deviation of the gradient of the critic (as a function of the network’s input) from one. We present theoretical arguments why using a weaker regularization term enforcing the Lipschitz constraint is preferable. These arguments are supported by experimental results on toy data sets.

    On the regularization of Wasserstein GANs
    by Henning Petzka, Asja Fischer, Denis Lukovnicov

You must be logged in to reply to this topic.