论文标题
RAF:使用极有限的查询对面部识别的递归对抗攻击
RAF: Recursive Adversarial Attacks on Face Recognition Using Extremely Limited Queries
论文作者
论文摘要
最近成功的对抗性攻击面部识别表明,尽管面部识别模型取得了显着的进步,但它们仍然远远远远远远落后于人类的感知和认可。它揭示了深度卷积神经网络(CNN)作为面部识别模型的最先进的构件,以针对对抗性示例,这可能会对安全系统造成某些后果。以前对基于梯度的对抗攻击进行了广泛的研究,并证明对面部识别模型取得了成功。但是,找到每张面部的优化扰动需要将大量查询提交到目标模型。在本文中,我们提出了使用自动面部扭曲对面部识别的递归对抗性攻击,该面部扭曲需要极有限的查询才能欺骗目标模型。翘曲功能不是随机的面部翘曲过程,而是应用于眉毛,鼻子,嘴唇等特定检测到的区域。我们评估了基于决策的黑色盒子攻击设置中提出方法的鲁棒性,在该设置中,攻击者无法访问模型参数和梯度,但是目标模型提供了硬贝尔的预测和置信度得分。
Recent successful adversarial attacks on face recognition show that, despite the remarkable progress of face recognition models, they are still far behind the human intelligence for perception and recognition. It reveals the vulnerability of deep convolutional neural networks (CNNs) as state-of-the-art building block for face recognition models against adversarial examples, which can cause certain consequences for secure systems. Gradient-based adversarial attacks are widely studied before and proved to be successful against face recognition models. However, finding the optimized perturbation per each face needs to submitting the significant number of queries to the target model. In this paper, we propose recursive adversarial attack on face recognition using automatic face warping which needs extremely limited number of queries to fool the target model. Instead of a random face warping procedure, the warping functions are applied on specific detected regions of face like eyebrows, nose, lips, etc. We evaluate the robustness of proposed method in the decision-based black-box attack setting, where the attackers have no access to the model parameters and gradients, but hard-label predictions and confidence scores are provided by the target model.