论文标题
F3A-GAN:带有生成对抗网络的面部动画的面部流动
F3A-GAN: Facial Flow for Face Animation with Generative Adversarial Networks
论文作者
论文摘要
面部动画旨在从一组有条件的面部运动驱动的单一源图像中综合连续的面部图像,旨在合成连续的面部图像。先前的作品主要是将面部运动建模为具有1D或2D表示的条件(例如,动作单元,情绪代码,地标),通常会导致在某些复杂的场景(例如连续产生和大量转化)中导致低质量的结果。为了解决这个问题,条件应该满足两个要求,即,保持运动信息保存和几何连续性。为此,我们提出了一种基于3D几何流动流的新颖表示,称为面部流,以表示任何姿势的人脸的自然运动。与以前的其他条件相比,提出的面部流井控制着面部的连续变化。之后,为了利用面部流进行面部编辑,我们构建了一个合成框架,该框架生成了有条件的面部流动的连续图像。为了充分利用面部流动的运动信息,设计了一个层次条件框架,旨在以层次结构的方式将图像和动作特征从图像和运动特征中提取的多尺度外观特征组合在一起。然后,该框架将多个融合功能解码回图像。实验结果证明了与其他最先进方法相比,我们方法的有效性。
Formulated as a conditional generation problem, face animation aims at synthesizing continuous face images from a single source image driven by a set of conditional face motion. Previous works mainly model the face motion as conditions with 1D or 2D representation (e.g., action units, emotion codes, landmark), which often leads to low-quality results in some complicated scenarios such as continuous generation and largepose transformation. To tackle this problem, the conditions are supposed to meet two requirements, i.e., motion information preserving and geometric continuity. To this end, we propose a novel representation based on a 3D geometric flow, termed facial flow, to represent the natural motion of the human face at any pose. Compared with other previous conditions, the proposed facial flow well controls the continuous changes to the face. After that, in order to utilize the facial flow for face editing, we build a synthesis framework generating continuous images with conditional facial flows. To fully take advantage of the motion information of facial flows, a hierarchical conditional framework is designed to combine the extracted multi-scale appearance features from images and motion features from flows in a hierarchical manner. The framework then decodes multiple fused features back to images progressively. Experimental results demonstrate the effectiveness of our method compared to other state-of-the-art methods.