As machine learning expanding application, there are more and more unignorable privacy and safety issues. Especially inference attacks against Machine Learning models allow adversaries to infer sensitive information about the target model, such as training data, model parameters, etc. Inference attacks can lead to serious consequences, including violating individuals privacy, compromising the intellectual property of the owner of the machine learning model. As far as concerned, researchers have studied and analyzed in depth several types of inference attacks, albeit in isolation, but there is still a lack of a holistic rick assessment of inference attacks against machine learning models, such as their application in different scenarios, the common factors affecting the performance of these attacks and the relationship among the attacks. As a result, this paper performs a holistic risk assessment of different inference attacks against Machine Learning models. This paper focuses on three kinds of representative attacks: membership inference attack, attribute inference attack and model stealing attack. And a threat model taxonomy is established. A total of 12 target models using three model architectures, including AlexNet, ResNet18 and Simple CNN, are trained on four datasets, namely CelebA, UTKFace, STL10 and FMNIST.
翻译:随着机器学习应用的扩大,越来越多的隐私和安全问题越来越不可忽视。特别是,对机器学习模型的攻击使对手能够推断出关于目标模型的敏感信息,如培训数据、模型参数等。推断攻击可能导致严重后果,包括侵犯个人隐私,损害机器学习模型所有者的知识产权。就相关而言,研究人员深入研究并深入分析了几种类型的推论攻击,尽管是孤立的,但是仍然缺乏对机器学习模型的推论性攻击的全面评估,例如这些模型在不同情景中的应用、影响这些攻击的进行的共同因素以及攻击之间的关系。因此,本文对针对机器学习模型的不同推论性攻击进行了全面风险评估。本文侧重于三类代表性攻击:成员推论攻击、推论攻击和盗窃模型攻击。还建立了威胁模型分类。使用三个模型结构的总共12个目标模型,包括AlexNet、ResNet18和简易CNN,在四个数据集上进行了培训,即CelibA、FKUT和STL。