论文标题
基于在线内核的生成对抗网络
Online Kernel based Generative Adversarial Networks
论文作者
论文摘要
在过去五年中,深度学习的主要突破之一是生成对抗网络(GAN),这是一种基于神经网络的生成模型,旨在模仿鉴于样本数据集的一些基本分布。与许多有监督的问题相反,在一对网络参数上,将GAN训练作为最小 - 最大问题提出了GAN训练,从而使GAN训练是最小化的。尽管从经验上讲,但在几个领域表现出了令人印象深刻的成功,但研究人员对不寻常的训练行为感到困惑,包括骑自行车所谓的模式崩溃。在本文中,我们首先提供了一种定量方法来探索GAN培训中的一些挑战,并从经验上展示了这与歧视者网络的参数性质的关系。我们提出了一种新颖的方法,可以通过依靠基于内核的非参数歧视器来解决许多此类问题,该歧视者高度适合在线培训---我们称此为基于在线内核的生成对抗网络(OKGAN)。我们从经验上表明,OKGAN减轻了许多培训问题,包括模式崩溃和骑自行车,并且更适合理论保证。相比,OKGAN在逆向KL差异方面比其他GAN公式在合成数据方面表现出色。在MNIST,SVHN和Celeba等古典视觉数据集上,表现出可比的性能。
One of the major breakthroughs in deep learning over the past five years has been the Generative Adversarial Network (GAN), a neural network-based generative model which aims to mimic some underlying distribution given a dataset of samples. In contrast to many supervised problems, where one tries to minimize a simple objective function of the parameters, GAN training is formulated as a min-max problem over a pair of network parameters. While empirically GANs have shown impressive success in several domains, researchers have been puzzled by unusual training behavior, including cycling so-called mode collapse. In this paper, we begin by providing a quantitative method to explore some of the challenges in GAN training, and we show empirically how this relates fundamentally to the parametric nature of the discriminator network. We propose a novel approach that resolves many of these issues by relying on a kernel-based non-parametric discriminator that is highly amenable to online training---we call this the Online Kernel-based Generative Adversarial Networks (OKGAN). We show empirically that OKGANs mitigate a number of training issues, including mode collapse and cycling, and are much more amenable to theoretical guarantees. OKGANs empirically perform dramatically better, with respect to reverse KL-divergence, than other GAN formulations on synthetic data; on classical vision datasets such as MNIST, SVHN, and CelebA, show comparable performance.