Federated learning (FL) has enabled global model training on decentralized data in a privacy-preserving way by aggregating model updates. However, for many natural language processing (NLP) tasks that utilize pre-trained language models (PLMs) with large numbers of parameters, there are considerable communication costs associated with FL. Recently, prompt tuning, which tunes some soft prompts without modifying PLMs, has achieved excellent performance as a new learning paradigm. Therefore we want to combine the two methods and explore the effect of prompt tuning under FL. In this paper, we propose "FedPrompt" as the first work study prompt tuning in a model split learning way using FL, and prove that split learning greatly reduces the communication cost, only 0.01% of the PLMs' parameters, with little decrease on accuracy both on IID and Non-IID data distribution. This improves the efficiency of FL method while also protecting the data privacy in prompt tuning.In addition, like PLMs, prompts are uploaded and downloaded between public platforms and personal users, so we try to figure out whether there is still a backdoor threat using only soft prompt in FL scenarios. We further conduct backdoor attacks by data poisoning on FedPrompt. Our experiments show that normal backdoor attack can not achieve a high attack success rate, proving the robustness of FedPrompt.We hope this work can promote the application of prompt in FL and raise the awareness of the possible security threats.
翻译:联邦学习联合会(FL)通过汇总模型更新,使得以保密方式对分散的数据进行全球示范培训成为了隐私保存方式的分散数据培训。然而,对于许多自然语言处理(NLP)任务而言,使用大量参数的经过预先训练的语言模型(PLM),与FL相关的通信成本相当。最近,快速调换(调试(调试一些软调试,但不修改PLM)作为新的学习范例,取得了出色的业绩。因此,我们希望将这两种方法结合起来,并探索在FL下迅速调试的效果。在本文中,我们提议“FedPrompt”作为第一份工作研究,用FL快速化的方法对模型的分解学习方式进行快速调整,并证明分解学习会极大地降低通信成本,只有PLM参数的0.01%,而IID和非IID数据分布的准确性几乎没有降低多少。这提高了FL方法的效率,同时也保护了快速调时的数据隐私。此外,公共平台和个人用户之间也上传和下载了快速调,因此我们试图找出在FPRO快速化攻击中是否还存在后门威胁。我们的安全性试验中的常规性试验,我们可以证明FPL成功率。