Federated learning (FL) has enabled global model training on decentralized data in a privacy-preserving way by aggregating model updates. However, for many natural language processing (NLP) tasks that utilize pre-trained language models (PLMs) with large numbers of parameters, there are considerable communication costs associated with FL. Recently, prompt tuning, which tunes some soft prompts without modifying PLMs, has achieved excellent performance as a new learning paradigm. Therefore we want to combine the two methods and explore the effect of prompt tuning under FL. In this paper, we propose "FedPrompt" to study prompt tuning in a model split aggregation way using FL, and prove that split aggregation greatly reduces the communication cost, only 0.01% of the PLMs' parameters, with little decrease on accuracy both on IID and Non-IID data distribution. This improves the efficiency of FL method while also protecting the data privacy in prompt tuning. In addition, like PLMs, prompts are uploaded and downloaded between public platforms and personal users, so we try to figure out whether there is still a backdoor threat using only soft prompts in FL scenarios. We further conduct backdoor attacks by data poisoning on FedPrompt. Our experiments show that normal backdoor attack can not achieve a high attack success rate, proving the robustness of FedPrompt. We hope this work can promote the application of prompt in FL and raise the awareness of the possible security threats.
翻译:联邦学习联合会(FL)通过汇总模型更新,使得以保密方式对分散的数据进行全球示范培训成为了以保密方式对分散的数据进行全球示范培训。然而,对于许多自然语言处理(NLP)任务而言,使用大量参数的经过预先训练的语言模型(PLM),与FL有关的通信成本相当。最近,快速调换(调试(调试)不修改PLM)使一些软调试(调试)(不修改PLM)取得了出色的业绩,作为一个新的学习范例。因此,我们希望将这两种方法结合起来,并探索在FL下迅速调时的效果。在本文中,我们提议“FedPPrompt”,以使用FL的模型分割方式研究快速调试算(调),以快速调试算(调试算)快速调(调试算出),以快速调(PL)的混合调(PL)参数的合并(PL)大幅降低通信成本,只有0.01%的参数,而ID和非II数据分布的准确性(调)数据分布很少降低。这提高了FL方法的效率,同时保护快速调调时保护数据。此外,还可以调(PROPL)程序可以上调时,还可以调(我们攻击的快速)在反向后测试(FM)的系统攻击)可能实现安全率的后测试(我们的安全率。