Federated Learning enables multiple data centers to train a central model collaboratively without exposing any confidential data. Even though deterministic models are capable of performing high prediction accuracy, their lack of calibration and capability to quantify uncertainty is problematic for safety-critical applications. Different from deterministic models, probabilistic models such as Bayesian neural networks are relatively well-calibrated and able to quantify uncertainty alongside their competitive prediction accuracy. Both of the approaches appear in the federated learning framework; however, the aggregation scheme of deterministic models cannot be directly applied to probabilistic models since weights correspond to distributions instead of point estimates. In this work, we study the effects of various aggregation schemes for variational Bayesian neural networks. With empirical results on three image classification datasets, we observe that the degree of spread for an aggregated distribution is a significant factor in the learning process. Hence, we present an investigation on the question of how to combine variational Bayesian networks in federated learning, while providing benchmarks for different aggregation settings.
翻译:联邦学习使多个数据中心能够在不暴露任何机密数据的情况下合作培训一个中央模型。即使确定性模型能够实现高预测准确性,但缺乏校准和量化不确定性的能力对于安全关键应用来说也是个问题。不同于确定性模型,巴伊西亚神经网络等概率模型相对比较有理,并且能够与竞争性预测准确性一道量化不确定性。这两种方法都出现在联邦学习框架中;然而,确定性模型的汇总计划不能直接适用于概率模型,因为重量与分布相对应,而不是点估计值。在这项工作中,我们研究了不同海湾神经网络的各种汇总计划的效果。关于三个图像分类数据集的经验结果,我们观察到,总体分布的普及程度是学习过程中的一个重要因素。因此,我们提出了如何在联邦学习过程中将阿拉伯湾网络的变异性结合的问题,同时为不同的汇总环境提供了基准。