论文标题
健忘症机器学习
Amnesiac Machine Learning
论文作者
论文摘要
被遗忘的权利是最近颁布的一般数据保护法规(GDPR)法律的一部分,该法律影响了任何具有欧盟居民数据的数据持有人。它使欧盟居民能够要求删除其个人数据,包括用于培训机器学习模型的培训记录。不幸的是,深度神经网络模型容易受到信息泄漏攻击的攻击,例如模型反转攻击,这些攻击从训练有素的模型和成员推理攻击中提取了类信息,这些攻击决定了模型培训数据中示例的存在。如果恶意政党能够进行攻击并学习要删除的私人信息,那么这意味着模特所有者没有正确保护用户的权利,并且他们的模型可能不符合GDPR法律。在本文中,我们提出了两种有效的方法,可以解决模型所有者或数据持有人如何从模型中删除个人数据的问题,以使它们可能不容易受到模型反演和成员推理攻击的影响,同时保持模型效率。我们首先提出一个现实世界中的威胁模型,该模型表明,简单地删除培训数据不足以保护用户。我们跟进两种删除数据的方法,即学习和失忆,使模型所有者能够保护自己免受此类攻击,同时遵守法规。我们提供了广泛的经验分析,表明这些方法确实是有效的,可以安全地应用的,有效地从训练有素的模型中删除了有关敏感数据的学习信息,同时保持模型疗效。
The Right to be Forgotten is part of the recently enacted General Data Protection Regulation (GDPR) law that affects any data holder that has data on European Union residents. It gives EU residents the ability to request deletion of their personal data, including training records used to train machine learning models. Unfortunately, Deep Neural Network models are vulnerable to information leaking attacks such as model inversion attacks which extract class information from a trained model and membership inference attacks which determine the presence of an example in a model's training data. If a malicious party can mount an attack and learn private information that was meant to be removed, then it implies that the model owner has not properly protected their user's rights and their models may not be compliant with the GDPR law. In this paper, we present two efficient methods that address this question of how a model owner or data holder may delete personal data from models in such a way that they may not be vulnerable to model inversion and membership inference attacks while maintaining model efficacy. We start by presenting a real-world threat model that shows that simply removing training data is insufficient to protect users. We follow that up with two data removal methods, namely Unlearning and Amnesiac Unlearning, that enable model owners to protect themselves against such attacks while being compliant with regulations. We provide extensive empirical analysis that show that these methods are indeed efficient, safe to apply, effectively remove learned information about sensitive data from trained models while maintaining model efficacy.