论文标题

质量公制的肖像肖像线绘图生成未配对的培训数据

Quality Metric Guided Portrait Line Drawing Generation from Unpaired Training Data

论文作者

Yi, Ran, Liu, Yong-Jin, Lai, Yu-Kun, Rosin, Paul L.

论文摘要

面部肖像线图是一种独特的艺术风格,具有高度抽象和表现力。但是,由于其高语义限制,许多现有方法学会使用配对的训练数据来生成肖像画,这是昂贵且耗时的。在本文中,我们提出了一种新颖的方法,可以使用具有两个新功能的未配对训练数据自动将面部照片转换为肖像画。也就是说,我们的方法可以(1)学习使用单个网络以多种样式生成高质量的肖像画,(2)以训练数据中看不见的“新样式”生成肖像画。为了实现这些好处,我们(1)提出了一种新颖的肖像画质量指标,该指标是从人类的看法中学到的,(2)引入质量损失,以指导网络产生更好的肖像画。我们观察到,由于照片和肖像绘图域之间的显着信息不平衡,因此现有的未配合翻译方法(例如Cyclegan)倾向于在整个图纸中嵌入无形的重建信息,从而导致缺少重要的面部特征。为了解决这个问题,我们提出了一个新型的不对称周期映射,该映射强制可见重建信息,并且仅嵌入所选面部区域。与重要面部区域的局部判别器一起,我们的方法还保留了生成的图纸中的所有重要面部特征。发电机解剖进一步解释说,我们的模型学会在绘图过程中纳入面部语义信息。包括用户研究在内的广泛实验表明,我们的模型表现优于最先进的方法。

Face portrait line drawing is a unique style of art which is highly abstract and expressive. However, due to its high semantic constraints, many existing methods learn to generate portrait drawings using paired training data, which is costly and time-consuming to obtain. In this paper, we propose a novel method to automatically transform face photos to portrait drawings using unpaired training data with two new features; i.e., our method can (1) learn to generate high quality portrait drawings in multiple styles using a single network and (2) generate portrait drawings in a "new style" unseen in the training data. To achieve these benefits, we (1) propose a novel quality metric for portrait drawings which is learned from human perception, and (2) introduce a quality loss to guide the network toward generating better looking portrait drawings. We observe that existing unpaired translation methods such as CycleGAN tend to embed invisible reconstruction information indiscriminately in the whole drawings due to significant information imbalance between the photo and portrait drawing domains, which leads to important facial features missing. To address this problem, we propose a novel asymmetric cycle mapping that enforces the reconstruction information to be visible and only embedded in the selected facial regions. Along with localized discriminators for important facial regions, our method well preserves all important facial features in the generated drawings. Generator dissection further explains that our model learns to incorporate face semantic information during drawing generation. Extensive experiments including a user study show that our model outperforms state-of-the-art methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源