1. Mode Collapse

Untitled

<aside> 🧑‍🏫 Mode collapse happens when the generator learns to fool the discriminator by producing examples from a single class from the whole training dataset

</aside>

2. Problems with BCE loss

$$ J(\theta) = - \frac{1}{m}\sum^{m}_{i=1}[y^{(i)}\log{h(x^{(i)},\theta)}+(1-y^{(i)})\log{(1-h(x^{(i)},\theta))}] $$

Untitled

2-1. BCE loss in GANs

❗This unbalanced training difficulty of the discriminator and Generator causes the vanishing gradient problem

Untitled

3. Earth Mover’s Distance

Untitled

Untitled

4. Wasserstein Loss

4-1. BCE Loss Simplified

$$ J(\theta) = - \frac{1}{m}\sum^{m}_{i=1}[y^{(i)}\log{h(x^{(i)},\theta)}+(1-y^{(i)})\log{(1-h(x^{(i)},\theta))}] $$

$$ \min_{G} \max_{D}V(D,G) = \mathbb{E}[\log D(x)] + \mathbb{E}[\log (1-D(G(z)))] $$