Deploying machine learning models in production may allow adversaries to infer sensitive information about training data. There is a vast literature analyzing different types of inference risks, ranging from membership inference to reconstruction attacks. Inspired by the success of games (i.e., probabilistic experiments) to study security properties in cryptography, some authors describe privacy inference risks in machine learning using a similar game-based style. However, adversary capabilities and goals are often stated in subtly different ways from one presentation to the other, which makes it hard to relate and compose results. In this paper, we present a game-based framework to systematize the body of knowledge on privacy inference risks in machine learning. We use this framework to (1) provide a unifying structure for definitions of inference risks, (2) formally establish known relations among definitions, and (3) to uncover hitherto unknown relations that would have been difficult to spot otherwise.
翻译:在生产中部署机器学习模型可能导致对培训数据的敏感信息进行推断。已有大量文献分析了不同类型的推断风险,从成员推断到重构攻击。受到在密码学中研究安全属性的游戏(即概率性实验)成功的启示,一些作者使用类似的基于游戏的风格描述了机器学习中的隐私推断风险。然而,从一个演示到另一个演示,攻击者能力和目标经常以微妙的不同方式陈述,这使得难以关联和组合结果。在本文中,我们提出了一个基于游戏的框架来系统化机器学习中隐私推断风险的知识体系。我们使用该框架来(1)为推断风险的定义提供一个统一的结构,(2)形式化已知的定义之间的关系,并(3)发掘此前未知的关系,否则很难发现。