WGAN—Understanding the Wasserstein Distance
Explore the concept of Wasserstein distance and its application in Wasserstein GANs to improve GAN training stability. Understand the limitations of traditional loss functions, the benefits of Earth Mover's Distance, and techniques like gradient penalty and weight clipping to optimize adversarial learning.
We'll cover the following...
GANs have been known to be hard to train, especially if we have tried to build one from scratch. In this lesson, we will talk about how to use a better distance measure to improve the training of GANs, namely, the Wasserstein GAN.
The groundwork for
Analyzing the problems with vanilla GAN loss
Let’s go over the commonly used loss functions for GANs:
, which is the vanilla form of GAN loss.
The experimental results have shown that these loss functions work well in several applications. However, let’s dig deep into these functions and see what could go wrong when they don’t work so well.
Step 1: Problems with the first loss function:
Assume that the generator network is trained, and we need to find an optimal discriminator network D. We have the following:
In this formula,