Vertical federated learning (VFL), which enables multiple enterprises possessing non-overlapped features to strengthen their machine learning models without disclosing their private data and model parameters, has received increasing attention lately. Similar to other machine learning algorithms, VFL suffers from fairness issues, i.e., the learned model may be unfairly discriminatory over the group with sensitive attributes. To tackle this problem, we propose a fair VFL framework in this work. First, we systematically formulate the problem of training fair models in VFL, where the learning task is modeled as a constrained optimization problem. To solve it in a federated manner, we consider its equivalent dual form and develop an asynchronous gradient coordinate-descent ascent algorithm, where each data party performs multiple parallelized local updates per communication round to effectively reduce the number of communication rounds. We prove that the algorithm finds a $\delta$-stationary point of the dual objective in $\mathcal{O}(\delta^{-4})$ communication rounds under mild conditions. Finally, extensive experiments on three benchmark datasets demonstrate the superior performance of our method in training fair models.
翻译:纵向联合学习(VFL)使拥有非过度特征的多个企业能够加强其机器学习模式而无需披露其私人数据和模型参数,这种纵向联合学习(VFL)最近受到越来越多的关注。与其他机器学习算法一样,VFL也存在公平问题,即学习模式可能不公平地歧视具有敏感属性的群体。为了解决这一问题,我们提议在这项工作中建立一个公平的VFL框架。首先,我们系统地在VFL中制定培训公平模式的问题,在这种模式中,学习任务以有限的优化问题为模式。为了以联合方式解决这一问题,我们考虑其等效的双重形式,并开发一个无同步的梯度协调-白度天化算法,即每个数据方在每轮通信中进行多次平行的本地更新,以有效减少通信周期的数量。我们证明,在 $\mathcal{O} (\delta ⁇ 4} 4} 4} 4}通信回合中发现双重目标的固定点。最后,在温条件下,对三个基准数据集进行了广泛的实验,展示我们的方法在培训公平模型中的优异性表现。