论文标题
视频框架插值的流量引导可变形补偿网络
Flow Guidance Deformable Compensation Network for Video Frame Interpolation
论文作者
论文摘要
在过去几年中,基于运动的视频框架插值(VFI)方法随着深度卷积网络的发展取得了显着的进步。尽管流程图估计的不准确性通常会危害其性能,尤其是在大型运动和遮挡的情况下。在本文中,我们提出了一个流量引导可变形补偿网络(FGDCN),以克服现有基于运动的方法的缺点。 FGDCN将帧采样过程分解为两个步骤:流步和变形步骤。具体而言,流程步骤利用粗到细节的流量估计网络直接估计中间流量并同时合成锚固框架。为了确保估计流量的准确性,在此步骤中共同采用了蒸馏损失和面向任务的损失。在第一步中学到的流动先验的指导下,变形步骤设计了金字塔可变形补偿网络,以补偿流程的缺失细节。另外,提出了金字塔损失,以监督图像和频域中的模型。实验结果表明,所提出的算法在具有较少参数的各种数据集上实现了出色的性能。
Motion-based video frame interpolation (VFI) methods have made remarkable progress with the development of deep convolutional networks over the past years. While their performance is often jeopardized by the inaccuracy of flow map estimation, especially in the case of large motion and occlusion. In this paper, we propose a flow guidance deformable compensation network (FGDCN) to overcome the drawbacks of existing motion-based methods. FGDCN decomposes the frame sampling process into two steps: a flow step and a deformation step. Specifically, the flow step utilizes a coarse-to-fine flow estimation network to directly estimate the intermediate flows and synthesizes an anchor frame simultaneously. To ensure the accuracy of the estimated flow, a distillation loss and a task-oriented loss are jointly employed in this step. Under the guidance of the flow priors learned in step one, the deformation step designs a pyramid deformable compensation network to compensate for the missing details of the flow step. In addition, a pyramid loss is proposed to supervise the model in both the image and frequency domain. Experimental results show that the proposed algorithm achieves excellent performance on various datasets with fewer parameters.