In vertical federated learning, two-party split learning has become an important topic and has found many applications in real business scenarios. However, how to prevent the participants' ground-truth labels from possible leakage is not well studied. In this paper, we consider answering this question in an imbalanced binary classification setting, a common case in online business applications. We first show that, norm attack, a simple method that uses the norm of the communicated gradients between the parties, can largely reveal the ground-truth labels from the participants. We then discuss several protection techniques to mitigate this issue. Among them, we have designed a principled approach that directly maximizes the worst-case error of label detection. This is proved to be more effective in countering norm attack and beyond. We experimentally demonstrate the competitiveness of our proposed method compared to several other baselines.
翻译:在纵向联合学习中,双方分解学习已成为一个重要议题,在实际商业情景中发现许多应用。然而,如何防止参与者的地面真实标签可能渗漏的问题没有得到很好研究。在本文中,我们考虑在不平衡的二进制分类设置中回答这个问题,这是在线商业应用的一个常见案例。我们首先表明,规范攻击这一使用双方之间传递梯度规范的简单方法,可以大致显示参与者提供的地面真实标签。我们然后讨论一些保护技术来缓解这一问题。我们设计了一种原则性方法,直接最大限度地增加标签检测中最坏的错误。这证明在打击常规攻击方面和以后更为有效。我们实验性地展示了我们拟议方法与其他几个基线相比的竞争力。