Membership inference attacks aim to infer whether a data record has been used to train a target model by observing its predictions. In sensitive domains such as healthcare, this can constitute a severe privacy violation. In this work we attempt to address the existing knowledge gap by conducting an exhaustive study of membership inference attacks and defences in the domain of semantic image segmentation. Our findings indicate that for certain threat models, these learning settings can be considerably more vulnerable than the previously considered classification settings. We additionally investigate a threat model where a dishonest adversary can perform model poisoning to aid their inference and evaluate the effects that these adaptations have on the success of membership inference attacks. We quantitatively evaluate the attacks on a number of popular model architectures across a variety of semantic segmentation tasks, demonstrating that membership inference attacks in this domain can achieve a high success rate and defending against them may result in unfavourable privacy-utility trade-offs or increased computational costs.
翻译:成员身份推断攻击旨在推断数据记录是否被用于通过观察预测来培训目标模型。在保健等敏感领域,这可能构成严重的侵犯隐私行为。在这项工作中,我们试图通过对成员身份推断攻击和语义图像分割领域的防御进行详尽无遗的研究来弥补现有的知识差距。我们的研究结果表明,对某些威胁模式而言,这些学习环境可能比先前考虑的分类环境要脆弱得多。我们还调查了一种威胁模型,其中不诚实的对手可以进行模型中毒,协助其推断,并评估这些调整对成员身份推断攻击的成功产生的影响。我们量化评估了对一系列流行模型结构的攻击,这些攻击跨越了各种语义分割任务,表明这一领域的成员身份推断攻击可以取得高成功率,并防范这些攻击可能导致不利的隐私效用交易或增加计算费用。