论文标题
当关系网络遇到gans时:与三胞胎损失的关系gan
When Relation Networks meet GANs: Relation GANs with Triplet Loss
论文作者
论文摘要
尽管最近的研究在使用生成的对抗网络(GAN)生成逼真的图像方面取得了显着进步,但缺乏训练稳定性仍然是大多数gan的持久关注,尤其是在高分辨率输入和复杂数据集上。由于随机生成的分布几乎无法与实际分布重叠,因此训练剂通常会遭受梯度消失的问题。已经提出了许多方法来通过使用经验技术来限制歧视者的能力,例如减肥,梯度惩罚,频谱归一化等。在本文中,我们提供了一种更有原则的方法,作为解决此问题的替代方法。我们没有通过训练歧视器来研究配对样品之间的关系,而不是训练歧视者以区分真实和假输入样本,以训练歧视器将配对样品与相同分布和不同分布的样本分开。为此,我们探索了歧视者的关系网络体系结构,并设计了一个三重态损失,该损失可以更好地概括和稳定性。基准数据集上的广泛实验表明,提出的关系歧视者和新损失可以对可变视觉任务(包括无条件和条件图像产生和图像翻译)提供重大改进。
Though recent research has achieved remarkable progress in generating realistic images with generative adversarial networks (GANs), the lack of training stability is still a lingering concern of most GANs, especially on high-resolution inputs and complex datasets. Since the randomly generated distribution can hardly overlap with the real distribution, training GANs often suffers from the gradient vanishing problem. A number of approaches have been proposed to address this issue by constraining the discriminator's capabilities using empirical techniques, like weight clipping, gradient penalty, spectral normalization etc. In this paper, we provide a more principled approach as an alternative solution to this issue. Instead of training the discriminator to distinguish real and fake input samples, we investigate the relationship between paired samples by training the discriminator to separate paired samples from the same distribution and those from different distributions. To this end, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability. Extensive experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks including unconditional and conditional image generation and image translation.