In this paper, we explore a new knowledge-amalgamation problem, termed Federated Selective Aggregation (FedSA). The goal of FedSA is to train a student model for a new task with the help of several decentralized teachers, whose pre-training tasks and data are different and agnostic. Our motivation for investigating such a problem setup stems from a recent dilemma of model sharing. Many researchers or institutes have spent enormous resources on training large and competent networks. Due to the privacy, security, or intellectual property issues, they are, however, not able to share their own pre-trained models, even if they wish to contribute to the community. The proposed FedSA offers a solution to this dilemma and makes it one step further since, again, the learned student may specialize in a new task different from all of the teachers. To this end, we proposed a dedicated strategy for handling FedSA. Specifically, our student-training process is driven by a novel saliency-based approach that adaptively selects teachers as the participants and integrates their representative capabilities into the student. To evaluate the effectiveness of FedSA, we conduct experiments on both single-task and multi-task settings. Experimental results demonstrate that FedSA effectively amalgamates knowledge from decentralized models and achieves competitive performance to centralized baselines.
翻译:在本文中,我们探索了一个新的知识放大问题,称为联邦选择性聚合(FedSA) 。FedSA的目标是在一些分散教师的帮助下,为新任务培训学生模式,这些教师的训练前任务和数据不同,而且不可想象。我们调查这一问题的动机来自最近一个模式分享的两难处境。许多研究人员或研究所将大量资源用于培训大型和胜任的网络。然而,由于隐私、安全或知识产权问题,他们无法分享他们自己经过培训的模型,即使他们愿意为社区作出贡献。拟议的FedSA提供了解决这一难题的办法,并使它更进一步,因为学习的学生可能专门从事与所有教师不同的新任务。为此目的,我们提出了专门处理FedSA的战略。具体地说,我们的学生培训进程受到基于新颖的突出方法的驱动,这种方法以适应方式挑选教师作为参与者,并将他们的代表能力纳入学生之中。为了评估FedSA的有效性,我们从FSA的单一任务和多功能实验模型到MFD-CA级模型的有效测试结果。