In vertical federated learning (FL), the features of a data sample are distributed across multiple agents. As such, inter-agent collaboration can be beneficial not only during the learning phase, as is the case for standard horizontal FL, but also during the inference phase. A fundamental theoretical question in this setting is how to quantify the cost, or performance loss, of decentralization for learning and/or inference. In this paper, we consider general supervised learning problems with any number of agents, and provide a novel information-theoretic quantification of the cost of decentralization in the presence of privacy constraints on inter-agent communication within a Bayesian framework. The cost of decentralization for learning and/or inference is shown to be quantified in terms of conditional mutual information terms involving features and label variables.
翻译:在纵向联合学习(FL)中,数据抽样的特点分布在多个代理商之间,因此,机构间协作不仅在学习阶段有益,如标准横向FL那样,而且在推断阶段也有益,在这一背景下,一个根本的理论问题是如何量化为学习和(或)推断而下放权力的成本或绩效损失。在本文件中,我们考虑与任何若干代理商存在普遍监督的学习问题,并在巴伊西亚框架内对代理人之间通信的隐私存在限制的情况下,对权力下放的成本进行新的信息理论量化。学习和(或)推断权力下放的成本以有条件的相互信息术语(包括特点和标签变量)加以量化。