Due to privacy concerns, multi-party gradient tree boosting algorithms have become widely popular amongst machine learning researchers and practitioners. However, limited existing works have focused on vertically partitioned datasets, and the few existing works are either not scalable or tend to leak information. Thus, in this work, we propose SSXGB which is a scalable and secure multi-party gradient tree boosting framework for vertically partitioned datasets with partially outsourced computations. Specifically, we employ an additive homomorphic encryption (HE) scheme for security. We design two sub-protocols based on the HE scheme to perform non-linear operations associated with gradient tree boosting algorithms. Next, we propose a secure training and a secure prediction algorithms under the SSXGB framework. Then we provide theoretical security and communication analysis for the proposed framework. Finally, we evaluate the performance of the framework with experiments using two real-world datasets.
翻译:由于隐私问题,多党梯度树增殖算法在机器学习研究人员和从业者中广为人知,然而,有限的现有工程侧重于垂直分割的数据集,少数现有工程要么无法缩放,要么容易泄漏信息。因此,在这项工作中,我们提议采用SSXGB,这是一个可缩放和安全的多党梯度树增殖框架,用于部分外包计算垂直分割的数据集。具体地说,我们采用添加式同系物加密(HE)安全办法。我们根据HE计划设计了两个子程序,以进行与梯度树增殖算法有关的非线性操作。我们随后提议在SSXGB框架下进行安全培训和安全预测算法。我们随后为拟议的框架提供理论安全和通信分析。最后,我们用两个真实世界数据集来试验框架的绩效。