On the regularization of Wasserstein GANs
This topic contains 0 replies, has 1 voice, and was last updated by arXiv 1 year, 3 months ago.

On the regularization of Wasserstein GANs
Since their invention, generative adversarial networks (GANs) have become a popular approach for learning to model a distribution of real (unlabeled) data. Convergence problems during training are overcome by Wasserstein GANs which minimize the distance between the model and the empirical distribution in terms of a different metric, but thereby introduce a Lipschitz constraint into the optimization problem. A simple way to enforce the Lipschitz constraint on the class of functions, which can be modeled by the neural network, is weight clipping. It was proposed that training can be improved by instead augmenting the loss by a regularization term that penalizes the deviation of the gradient of the critic (as a function of the network’s input) from one. We present theoretical arguments why using a weaker regularization term enforcing the Lipschitz constraint is preferable. These arguments are supported by experimental results on toy data sets.
On the regularization of Wasserstein GANs
by Henning Petzka, Asja Fischer, Denis Lukovnicov
https://arxiv.org/pdf/1709.08894v1.pdf
You must be logged in to reply to this topic.