论文标题
限制黑盒对抗性攻击对Deepfake面部交换
Restricted Black-box Adversarial Attack Against DeepFake Face Swapping
论文作者
论文摘要
DeepFake面部交换对在线安全和社交媒体构成了重大威胁,这些威胁可以用一个完全不同的人的目标面对任意照片/视频中的来源面孔。为了防止这种欺诈行为,一些研究人员已经开始研究针对Deepfake或面对操纵的对抗方法。但是,现有作品着眼于白色框设置或由大量查询驱动的黑框设置,这严重限制了这些方法的实际应用。为了解决这个问题,我们引入了实用的对抗攻击,该攻击不需要对面部图像伪造模型的任何查询。我们的方法建立在替代模型上,以实现面部重建,然后将对抗性示例从替代模型直接传输到无法访问的黑盒深击模型。特别是,我们提出了可转移的周期对手生成对抗网络(TCA-GAN),以构建对抗扰动,以破坏未知的深泡系统。我们还提出了一个新型的调查后模块,以增强产生的对抗性实例的可转移性。为了全面衡量我们的方法的有效性,我们为未来发展构建了深层对抗攻击的挑战性基准。广泛的实验令人印象深刻地表明,所提出的对抗攻击方法使深击面部图像的视觉质量下降,从而更容易被人类和算法检测到它们。此外,我们证明了所提出的算法可以推广,以针对各种面部翻译方法提供面部图像保护。
DeepFake face swapping presents a significant threat to online security and social media, which can replace the source face in an arbitrary photo/video with the target face of an entirely different person. In order to prevent this fraud, some researchers have begun to study the adversarial methods against DeepFake or face manipulation. However, existing works focus on the white-box setting or the black-box setting driven by abundant queries, which severely limits the practical application of these methods. To tackle this problem, we introduce a practical adversarial attack that does not require any queries to the facial image forgery model. Our method is built on a substitute model persuing for face reconstruction and then transfers adversarial examples from the substitute model directly to inaccessible black-box DeepFake models. Specially, we propose the Transferable Cycle Adversary Generative Adversarial Network (TCA-GAN) to construct the adversarial perturbation for disrupting unknown DeepFake systems. We also present a novel post-regularization module for enhancing the transferability of generated adversarial examples. To comprehensively measure the effectiveness of our approaches, we construct a challenging benchmark of DeepFake adversarial attacks for future development. Extensive experiments impressively show that the proposed adversarial attack method makes the visual quality of DeepFake face images plummet so that they are easier to be detected by humans and algorithms. Moreover, we demonstrate that the proposed algorithm can be generalized to offer face image protection against various face translation methods.