论文标题
通过Langevin Stein变化梯度下降稳定生成对抗网的训练
Stabilizing Training of Generative Adversarial Nets via Langevin Stein Variational Gradient Descent
论文作者
论文摘要
但是,以学习复杂数据分配的能力而闻名的生成对抗网络(GAN)在训练过程中很棘手,这可能会导致模式崩溃或性能恶化。目前处理GAN问题的方法几乎利用了一些实用的培训技术来实现正规化,另一方面,这破坏了GAN的融合和理论声音。在本文中,我们建议通过基于粒子的新变异推断(Langevin Stein变化梯度下降(LSVGD))稳定GAN训练,这不仅继承了原始SVGD的灵活性和效率,而且旨在通过将其不稳定问题纳入更新动力学中,以解决其不稳定性问题。我们进一步证明,通过正确调整噪声方差,LSVGD模拟了一个固定分布正好是目标分布的langevin过程。我们还表明,LSVGD动力学具有隐式正则化,能够增强粒子的扩散和多样性。最后,无论采用哪种损失函数,我们都提出了一种对一般GAN训练程序应用基于粒子的变异推断的有效方法。一个合成数据集和三个流行基准数据集的实验结果-CIFAR-10,Tiny-Imagenet和Celeba验证LSVGD可以显着提高各种GAN模型的性能和稳定性。
Generative adversarial networks (GANs), famous for the capability of learning complex underlying data distribution, are however known to be tricky in the training process, which would probably result in mode collapse or performance deterioration. Current approaches of dealing with GANs' issues almost utilize some practical training techniques for the purpose of regularization, which on the other hand undermines the convergence and theoretical soundness of GAN. In this paper, we propose to stabilize GAN training via a novel particle-based variational inference -- Langevin Stein variational gradient descent (LSVGD), which not only inherits the flexibility and efficiency of original SVGD but aims to address its instability issues by incorporating an extra disturbance into the update dynamics. We further demonstrate that by properly adjusting the noise variance, LSVGD simulates a Langevin process whose stationary distribution is exactly the target distribution. We also show that LSVGD dynamics has an implicit regularization which is able to enhance particles' spread-out and diversity. At last we present an efficient way of applying particle-based variational inference on a general GAN training procedure no matter what loss function is adopted. Experimental results on one synthetic dataset and three popular benchmark datasets -- Cifar-10, Tiny-ImageNet and CelebA validate that LSVGD can remarkably improve the performance and stability of various GAN models.