The increasing concerns about data privacy and security drives the emergence of a new field of studying privacy-preserving machine learning from isolated data sources, i.e., \textit{federated learning}. Vertical federated learning, where different parties hold different features for common users, has a great potential of driving a more variety of business cooperation among enterprises in different fields. Decision tree models especially decision tree ensembles are a class of widely applied powerful machine learning models with high interpretability and modeling efficiency. However, the interpretability are compromised in these works such as SecureBoost since the feature names are not exposed to avoid possible data breaches due to the unprotected decision path. In this paper, we shall propose Fed-EINI, an efficient and interpretable inference framework for federated decision tree models with only one round of multi-party communication. We shall compute the candidate sets of leaf nodes based on the local data at each party in parallel, followed by securely computing the weight of the only leaf node in the intersection of the candidate sets. We propose to protect the decision path by the efficient additively homomorphic encryption method, which allows the disclosure of feature names and thus makes the federated decision trees interpretable. The advantages of Fed-EINI will be demonstrated through theoretical analysis and extensive numerical results. Experiments show that the inference efficiency is improved by over $50\%$ in average.
翻译:对数据隐私和安全的日益关切促使出现了从孤立的数据来源(即\ textit{federate learning})学习隐私保存机器的新领域。 纵向联合学习,即不同当事方对共同用户具有不同的特征,极有可能推动不同领域的企业之间更为多样的商业合作。决策树模型,特别是决策树集合,是广泛应用的强大机器学习模型的一类,可解释性和建模效率很高。但是,在Seet Boost等工程中,可解释性受到损害,因为特征名称不会暴露于避免因不受保护的决策路径而可能发生的数据破损。在本文件中,我们将提出Fed-EINI,这是一个高效和可解释的推论框架,即对Federd-EINI, 是一个高效和可解释的决定树模型框架,只有一轮多党交流。我们应该根据每个当事方的当地数据,同时对叶节点数进行计算,然后安全地计算候选人组合中唯一叶节点的重量。我们提议通过高效的同系定式加密路径来保护决策路径。我们提议,通过对美化的同系性加密方法进行高效的精确化化化化化化的对结果进行解释,这样通过实验性化的实验性分析,可以展示。