As predictive models are increasingly being employed to make consequential decisions, there is a growing emphasis on developing techniques that can provide algorithmic recourse to affected individuals. While such recourses can be immensely beneficial to affected individuals, potential adversaries could also exploit these recourses to compromise privacy. In this work, we make the first attempt at investigating if and how an adversary can leverage recourses to infer private information about the underlying model's training data. To this end, we propose a series of novel membership inference attacks which leverage algorithmic recourse. More specifically, we extend the prior literature on membership inference attacks to the recourse setting by leveraging the distances between data instances and their corresponding counterfactuals output by state-of-the-art recourse methods. Extensive experimentation with real world and synthetic datasets demonstrates significant privacy leakage through recourses. Our work establishes unintended privacy leakage as an important risk in the widespread adoption of recourse methods.
翻译:随着预测模型越来越多地被用于做出相应的决定,人们越来越强调开发能够向受影响个人提供算法追索的技术,虽然这种追索可能给受影响个人带来巨大好处,但潜在的对手也可以利用这些求助来损害隐私。在这项工作中,我们首先试图调查对手是否以及如何利用推断私人信息的方法来推断基本模型的培训数据。为此,我们提出一系列利用算法追索的新的成员推论攻击。更具体地说,我们把先前关于成员推论攻击的文献扩大到追索环境,办法是利用最新追索方法在数据案例与其对应的反事实产出之间的距离。对真实世界和合成数据集的广泛实验表明通过追索方式大量隐私泄漏。我们的工作将意外隐私渗漏确定为广泛采用追索方法的一个重要风险。