Kim Sehwan, Song Qifan, Liang Faming
Department of Statistics, Purdue University, West Lafayette, IN 47907.
Stat Sin. 2025 Apr;35(2):897-918. doi: 10.5705/ss.202022.0404.
The Generative Adversarial Network (GAN) was recently introduced in the literature as a novel machine learning method for training generative models. It has many applications in statistics such as nonparametric clustering and nonparametric conditional independence tests. However, training the GAN is notoriously difficult due to the issue of mode collapse, which refers to the lack of diversity among generated data. In this paper, we identify the reasons why the GAN suffers from this issue, and to address it, we propose a new formulation for the GAN based on randomized decision rules. In the new formulation, the discriminator converges to a fixed point while the generator converges to a distribution at the Nash equilibrium. We propose to train the GAN by an empirical Bayes-like method by treating the discriminator as a hyper-parameter of the posterior distribution of the generator. Specifically, we simulate generators from its posterior distribution conditioned on the discriminator using a stochastic gradient Markov chain Monte Carlo (MCMC) algorithm, and update the discriminator using stochastic gradient descent along with simulations of the generators. We establish convergence of the proposed method to the Nash equilibrium. Apart from image generation, we apply the proposed method to nonparametric clustering and nonparametric conditional independence tests. A portion of the numerical results is presented in the supplementary material.
生成对抗网络(GAN)最近作为一种用于训练生成模型的新型机器学习方法被引入文献。它在统计学中有许多应用,如非参数聚类和非参数条件独立性检验。然而,由于模式崩溃问题,训练GAN非常困难,模式崩溃是指生成数据缺乏多样性。在本文中,我们确定了GAN出现此问题的原因,并为解决该问题,基于随机决策规则提出了一种新的GAN公式。在新公式中,判别器收敛到一个不动点,而生成器在纳什均衡处收敛到一个分布。我们建议通过一种类似经验贝叶斯的方法训练GAN,将判别器视为生成器后验分布的超参数。具体来说,我们使用随机梯度马尔可夫链蒙特卡罗(MCMC)算法从基于判别器的后验分布中模拟生成器,并使用随机梯度下降以及生成器的模拟来更新判别器。我们证明了所提出方法收敛到纳什均衡。除了图像生成,我们还将所提出的方法应用于非参数聚类和非参数条件独立性检验。部分数值结果在补充材料中给出。