论文标题
有效的属性学习:选择性从特征表示中选择性删除输入属性
Efficient Attribute Unlearning: Towards Selective Removal of Input Attributes from Feature Representations
论文作者
论文摘要
最近,《隐私法规》的制定促进了机器未学习范式的兴起。现有的对机器学习的研究主要集中于样本的学习,因此,学到的模型不会在样本级别揭示用户的隐私。然而,我们认为,这种选择性去除的能力也应在属性级别上呈现,尤其是对于与主要任务无关的属性,例如,在面部识别系统中认可的人是否戴着眼镜或该人的年龄范围。通过全面的文献综述,发现现有关于属性相关问题(如公平和偏见的学习)的研究无法正确解决上述问题。为了弥合此差距,我们提出了一个从特征表示中选择性删除输入属性的范式,我们将其命名为“属性删除”。在此范式中,根据培训阶段,将准确地捕获某些属性,并根据其相互信息从训练阶段脱离。将逐渐消除特定属性,以及辅助培训程序,而与主要任务相关的其余属性则用于实现竞争性模型性能。考虑到训练过程中的计算复杂性,我们不仅提供了一种理论上近似培训方法,而且提出了加速训练过程的加速方案。我们通过跨越几个数据集和模型来验证我们的方法,并证明我们的设计可以保持模型保真度并以高效率达到盛行的学习效率。拟议的未学习范式为将来的机器学习系统奠定了基础,并将成为最新的与隐私有关的立法的重要组成部分。
Recently, the enactment of privacy regulations has promoted the rise of the machine unlearning paradigm. Existing studies of machine unlearning mainly focus on sample-wise unlearning, such that a learnt model will not expose user's privacy at the sample level. Yet we argue that such ability of selective removal should also be presented at the attribute level, especially for the attributes irrelevant to the main task, e.g., whether a person recognized in a face recognition system wears glasses or the age range of that person. Through a comprehensive literature review, it is found that existing studies on attribute-related problems like fairness and de-biasing learning cannot address the above concerns properly. To bridge this gap, we propose a paradigm of selectively removing input attributes from feature representations which we name `attribute unlearning'. In this paradigm, certain attributes will be accurately captured and detached from the learned feature representations at the stage of training, according to their mutual information. The particular attributes will be progressively eliminated along with the training procedure towards convergence, while the rest of attributes related to the main task are preserved for achieving competitive model performance. Considering the computational complexity during the training process, we not only give a theoretically approximate training method, but also propose an acceleration scheme to speed up the training process. We validate our method by spanning several datasets and models and demonstrate that our design can preserve model fidelity and reach prevailing unlearning efficacy with high efficiency. The proposed unlearning paradigm builds a foundation for future machine unlearning system and will become an essential component of the latest privacy-related legislation.