论文标题
小波知识蒸馏:迈向有效的图像到图像翻译
Wavelet Knowledge Distillation: Towards Efficient Image-to-Image Translation
论文作者
论文摘要
图像到图像翻译中的生成对抗网络(GAN)已取得了显着的成就。但是,由于大量参数,最先进的gan通常会遭受低效率和庞大的记忆使用量。为了应对这一挑战,首先,本文从频率的角度研究了gan的性能。结果表明,甘恩(Gans),尤其是小甘斯(Gans)缺乏产生高质量高频信息的能力。为了解决这个问题,我们提出了一种新的知识蒸馏方法,称为小波知识蒸馏。小波知识蒸馏没有直接蒸馏出教师的生成图像,而是首先将图像分解为具有离散小波转换的不同频段,然后仅将高频带。结果,学生甘可以更多地关注其在高频带上的学习。实验表明,我们的方法导致压缩的7.08倍,并且在Cyclean上加速了6.80倍,几乎没有性能下降。此外,我们研究了鉴别器和发电机之间的关系,这些关系表明歧视器的压缩可以促进压缩发电机的性能。
Remarkable achievements have been attained with Generative Adversarial Networks (GANs) in image-to-image translation. However, due to a tremendous amount of parameters, state-of-the-art GANs usually suffer from low efficiency and bulky memory usage. To tackle this challenge, firstly, this paper investigates GANs performance from a frequency perspective. The results show that GANs, especially small GANs lack the ability to generate high-quality high frequency information. To address this problem, we propose a novel knowledge distillation method referred to as wavelet knowledge distillation. Instead of directly distilling the generated images of teachers, wavelet knowledge distillation first decomposes the images into different frequency bands with discrete wavelet transformation and then only distills the high frequency bands. As a result, the student GAN can pay more attention to its learning on high frequency bands. Experiments demonstrate that our method leads to 7.08 times compression and 6.80 times acceleration on CycleGAN with almost no performance drop. Additionally, we have studied the relation between discriminators and generators which shows that the compression of discriminators can promote the performance of compressed generators.