论文标题
概率分类对抗性攻击和对抗训练
Probabilistic Categorical Adversarial Attack & Adversarial Training
论文作者
论文摘要
对抗性例子的存在给人们带来了深层神经网络(DNN)在安全至关重要的任务中的巨大关注。但是,如何使用分类数据产生对抗性示例是一个重要的问题,但缺乏广泛的探索。以前建立的方法利用贪婪的搜索方法,这可能非常耗时,以进行成功的攻击。这也限制了对对抗性训练的发展和分类数据的潜在防御能力。为了解决这个问题,我们提出了概率分类对抗攻击(PCAA),该攻击将离散的优化问题转移到连续问题,该问题可以通过投影梯度下降有效地解决。在我们的论文中,我们理论上分析了其最佳和时间复杂性,以证明其比目前基于贪婪的攻击的重要优势。此外,根据我们的攻击,我们提出了一个有效的对抗训练框架。通过一项全面的实证研究,我们证明了提议的攻击和防御算法的有效性。
The existence of adversarial examples brings huge concern for people to apply Deep Neural Networks (DNNs) in safety-critical tasks. However, how to generate adversarial examples with categorical data is an important problem but lack of extensive exploration. Previously established methods leverage greedy search method, which can be very time-consuming to conduct successful attack. This also limits the development of adversarial training and potential defenses for categorical data. To tackle this problem, we propose Probabilistic Categorical Adversarial Attack (PCAA), which transfers the discrete optimization problem to a continuous problem that can be solved efficiently by Projected Gradient Descent. In our paper, we theoretically analyze its optimality and time complexity to demonstrate its significant advantage over current greedy based attacks. Moreover, based on our attack, we propose an efficient adversarial training framework. Through a comprehensive empirical study, we justify the effectiveness of our proposed attack and defense algorithms.