Recently, the enactment of privacy regulations has promoted the rise of the machine unlearning paradigm. Existing studies of machine unlearning mainly focus on sample-wise unlearning, such that a learnt model will not expose user's privacy at the sample level. Yet we argue that such ability of selective removal should also be presented at the attribute level, especially for the attributes irrelevant to the main task, e.g., whether a person recognized in a face recognition system wears glasses or the age range of that person. Through a comprehensive literature review, it is found that existing studies on attribute-related problems like fairness and de-biasing learning cannot address the above concerns properly. To bridge this gap, we propose a paradigm of selectively removing input attributes from feature representations which we name `attribute unlearning'. In this paradigm, certain attributes will be accurately captured and detached from the learned feature representations at the stage of training, according to their mutual information. The particular attributes will be progressively eliminated along with the training procedure towards convergence, while the rest of attributes related to the main task are preserved for achieving competitive model performance. Considering the computational complexity during the training process, we not only give a theoretically approximate training method, but also propose an acceleration scheme to speed up the training process. We validate our method by spanning several datasets and models and demonstrate that our design can preserve model fidelity and reach prevailing unlearning efficacy with high efficiency. The proposed unlearning paradigm builds a foundation for future machine unlearning system and will become an essential component of the latest privacy-related legislation.
翻译:最近,隐私条例的颁布促进了机器不学习范式的兴起。现有的机器不学习研究主要侧重于抽样学不学,因此,学习的模型不会暴露用户在抽样一级的隐私。然而,我们争辩说,这种选择性删除的能力也应在属性层面提出,特别是对于与主要任务无关的属性而言,例如,在面对面识别系统中被承认的人是否戴眼镜或该人的年龄范围。通过综合文献审查,发现关于公平性和降低偏见学习等属性相关问题的现有研究无法正确解决上述关切。为了缩小这一差距,我们建议一种有选择地从我们称之为“归属不学习”的特征展示中去除输入属性的模式。在这一模式中,某些属性将准确地被捕获,并脱离在培训阶段学到的特征表现,例如,在培训程序的同时将逐渐消除特定属性,走向趋同,同时保留与主要任务相关的属性,以实现竞争性示范性业绩。考虑到在培训过程中的计算性复杂性不足,我们不仅提出有选择地从我们称之为“归属不学习”的特征上删除输入属性属性。在这个模式中,我们还提出一个只是从理论上接近性地展示我们的最新培训方法,并且提出一种更新我们当前培训方法。