论文标题
会员资格推断攻击通过对抗性示例
Membership Inference Attacks via Adversarial Examples
论文作者
论文摘要
机器学习和深度学习的提升导致了几个领域的显着改善。计算能力的急剧上升和大型数据集的集合都支持了这一变化。这样的庞大数据集通常包括可能代表隐私威胁的个人数据。会员推理攻击是一个新的研究方向,旨在恢复学习算法使用的培训数据。在本文中,我们开发了一种均值,以衡量利用数量的训练数据的泄漏,该数量是训练样本附近训练有素模型的总变化。我们通过提供一种新颖的防御机制来扩展工作。通过说服数值实验,经验证据支持我们的贡献。
The raise of machine learning and deep learning led to significant improvement in several domains. This change is supported by both the dramatic rise in computation power and the collection of large datasets. Such massive datasets often include personal data which can represent a threat to privacy. Membership inference attacks are a novel direction of research which aims at recovering training data used by a learning algorithm. In this paper, we develop a mean to measure the leakage of training data leveraging a quantity appearing as a proxy of the total variation of a trained model near its training samples. We extend our work by providing a novel defense mechanism. Our contributions are supported by empirical evidence through convincing numerical experiments.