Johnson Rie, Zhang Tong
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):17-32. doi: 10.1109/TPAMI.2019.2924428. Epub 2020 Dec 4.
Generative adversarial networks (GAN) are trained through a minimax game between a generator and a discriminator to generate data that mimics observations. While being widely used, GAN training is known to be empirically unstable. This paper presents a new theory for generative adversarial methods that does not rely on the traditional minimax formulation. Our theory shows that with a strong discriminator, a good generator can be obtained by composite functional gradient learning, so that several distance measures (including the KL divergence and the JS divergence) between the probability distributions of real data and generated data are simultaneously improved after each functional gradient step until converging to zero. This new point of view leads to stable procedures for training generative models. It also gives a new theoretical insight into the original GAN. Empirical results on image generation show the effectiveness of our new method.
生成对抗网络(GAN)通过生成器和判别器之间的极小极大博弈进行训练,以生成模仿观测值的数据。虽然GAN被广泛使用,但经验表明GAN训练是不稳定的。本文提出了一种新的生成对抗方法理论,该理论不依赖于传统的极小极大公式。我们的理论表明,在判别器很强的情况下,可以通过复合函数梯度学习获得一个好的生成器,这样在每次函数梯度步骤之后,真实数据和生成数据的概率分布之间的几种距离度量(包括KL散度和JS散度)会同时得到改善,直至收敛到零。这种新观点带来了训练生成模型的稳定方法。它还为原始GAN提供了新的理论见解。图像生成的实证结果表明了我们新方法的有效性。