论文标题
通过对抗性鲁棒性评估会员推断
Evaluating Membership Inference Through Adversarial Robustness
论文作者
论文摘要
在许多应用中,深度学习的使用正在升级。由于其出色的性能,除了传统的应用程序外,它还用于各种安全性和隐私敏感区域。深度学习功效的关键方面之一是拥有丰富的数据。这种特征导致数据使用可能是高度敏感和私人的,这反过来又引起了公众深入学习的警惕。会员推理攻击被认为是致命的,因为它们可以用来弄清楚一块数据是否属于培训数据集。对于培训数据信息及其特征的泄漏,这可能是有问题的。为了强调这些类型的攻击的重要性,我们通过通过在白色盒子设置下的标签平滑来调整对抗性扰动的方向,提出了一种基于对抗性鲁棒性的成员推理攻击的增强方法。我们在三个数据集上评估了我们提出的方法:Fashion-Mnist,CIFAR-10和CIFAR-100。我们的实验结果表明,在攻击正常训练的模型时,我们方法的性能超过了现有的基于对抗性鲁棒性的方法。此外,通过将我们的技术与最先进的成员推理方法进行比较,我们提出的方法在攻击受对抗训练的模型时还显示出更好的性能。重现此工作结果的代码可在\ url {https://github.com/plll4zzx/evaluating-membership-inference-inference-inference-infere--though-though-versar-versarial-robustness}中获得。
The usage of deep learning is being escalated in many applications. Due to its outstanding performance, it is being used in a variety of security and privacy-sensitive areas in addition to conventional applications. One of the key aspects of deep learning efficacy is to have abundant data. This trait leads to the usage of data which can be highly sensitive and private, which in turn causes wariness with regard to deep learning in the general public. Membership inference attacks are considered lethal as they can be used to figure out whether a piece of data belongs to the training dataset or not. This can be problematic with regards to leakage of training data information and its characteristics. To highlight the significance of these types of attacks, we propose an enhanced methodology for membership inference attacks based on adversarial robustness, by adjusting the directions of adversarial perturbations through label smoothing under a white-box setting. We evaluate our proposed method on three datasets: Fashion-MNIST, CIFAR-10, and CIFAR-100. Our experimental results reveal that the performance of our method surpasses that of the existing adversarial robustness-based method when attacking normally trained models. Additionally, through comparing our technique with the state-of-the-art metric-based membership inference methods, our proposed method also shows better performance when attacking adversarially trained models. The code for reproducing the results of this work is available at \url{https://github.com/plll4zzx/Evaluating-Membership-Inference-Through-Adversarial-Robustness}.