Revoking personal private data is one of the basic human rights, which has already been sheltered by several privacy-preserving laws in many countries. However, with the development of data science, machine learning and deep learning techniques, this right is usually neglected or violated as more and more patients' data are being collected and used for model training, especially in intelligent healthcare, thus making intelligent healthcare a sector where technology must meet the law, regulations, and privacy principles to ensure that the innovation is for the common good. In order to secure patients' right to be forgotten, we proposed a novel solution by using auditing to guide the forgetting process, where auditing means determining whether a dataset has been used to train the model and forgetting requires the information of a query dataset to be forgotten from the target model. We unified these two tasks by introducing a new approach called knowledge purification. To implement our solution, we developed AFS, a unified open-source software, which is able to evaluate and revoke patients' private data from pre-trained deep learning models. We demonstrated the generality of AFS by applying it to four tasks on different datasets with various data sizes and architectures of deep learning networks. The software is publicly available at \url{https://github.com/JoshuaChou2018/AFS}.
翻译:个人私人数据是基本人权之一,在许多国家,个人私人数据被若干保护隐私的法律所保护。然而,随着数据科学、机器学习和深层学习技术的发展,随着越来越多的病人数据被收集并用于示范培训,特别是在智能保健方面,使智能保健成为技术必须符合法律、条例和隐私原则以确保创新符合共同利益的一个部门,因此这项权利通常被忽视或受到侵犯。为了确保病人被遗忘的权利,我们提出了一个新的解决办法,即通过审计来指导遗忘过程,审计意味着确定是否使用了数据集来培训模型,而忘记需要查询数据集的信息才能从目标模型中遗忘。我们通过采用称为知识净化的新方法将这两项任务统一起来。为了实施我们的解决办法,我们开发了AFS, 这是一种统一的开放源软件,它能够从预先培训的深层学习模型中评估和撤销病人的私人数据。我们通过应用AFS20S,将它应用到不同数据大小和深层次学习网络结构的不同数据集的4项任务上,我们展示了AFSFSFS20 。