Federated Learning (FL) has emerged as a promising distributed learning paradigm with an added advantage of data privacy. With the growing interest in having collaboration among data owners, FL has gained significant attention of organizations. The idea of FL is to enable collaborating participants train machine learning (ML) models on decentralized data without breaching privacy. In simpler words, federated learning is the approach of ``bringing the model to the data, instead of bringing the data to the mode''. Federated learning, when applied to data which is partitioned vertically across participants, is able to build a complete ML model by combining local models trained only using the data with distinct features at the local sites. This architecture of FL is referred to as vertical federated learning (VFL), which differs from the conventional FL on horizontally partitioned data. As VFL is different from conventional FL, it comes with its own issues and challenges. In this paper, we present a structured literature review discussing the state-of-the-art approaches in VFL. Additionally, the literature review highlights the existing solutions to challenges in VFL and provides potential research directions in this domain.
翻译:联邦学习联合会(FL)已经成为一个充满希望的分布式学习模式,并增加了数据隐私的优势。随着对数据拥有者之间合作的兴趣日益浓厚,FL得到了各组织的极大关注。FL的想法是使合作参与者能够在不侵犯隐私的情况下对分散的数据进行机器学习模式的培训。更简单的说,联邦学习是“将模型引入数据,而不是将数据引入模式”的方法。当将数据垂直分布在参与者之间时,联邦学习能够建立一个完整的ML模型,将仅使用当地有不同特征的数据培训的地方模型结合起来。FLL的这一结构被称为纵向联合学习(VFL),不同于横向分割数据的传统FL。VLL不同于常规的FL,它涉及其自身的问题和挑战。在本文中,我们提出了一个结构化的文献审查,讨论VLFL中的最新方法。此外,文献审查强调了目前对VLF的挑战的解决办法,并提供了该领域的潜在研究方向。