Proposed as a solution to mitigate the privacy implications related to the adoption of deep learning solutions, Federated Learning (FL) enables large numbers of participants to successfully train deep neural networks without having to reveal the actual private training data. To date, a substantial amount of research has investigated the security and privacy properties of FL, resulting in a plethora of innovative attack and defense strategies. This paper thoroughly investigates the communication capabilities of an FL scheme. In particular, we show that a party involved in the FL learning process can use FL as a covert communication medium to send an arbitrary message. We introduce FedComm, a novel covert-communication technique that enables robust sharing and transfer of targeted payloads within the FL framework. Our extensive theoretical and empirical evaluations show that FedComm provides a stealthy communication channel, with minimal disruptions to the training process. Our experiments show that FedComm, allowed us to successfully deliver 100% of a payload in the order of kilobits before the FL procedure converges. Our evaluation also shows that FedComm is independent of the application domain and the neural network architecture used by the underlying FL scheme.
翻译:联邦学习联合会(FL)让大量参与者能够成功培训深神经网络,而不必披露实际的私人培训数据。迄今为止,大量研究已经调查了FL的安全和隐私性质,导致了大量的创新攻击和防御战略。本文彻底调查了FL计划的通信能力。特别是,我们表明,参与FL学习进程的一方可以将FL作为传递任意信息的秘密通信媒介。我们引入了FDComm,这是一种新的秘密通信技术,能够在FL框架内强有力地分享和转让目标有效载荷。我们广泛的理论和经验评估表明,FDComm提供了一个隐性通信渠道,对培训过程的干扰最小。我们的实验表明,FDComm允许我们在FL程序趋同之前以千比特的速度成功交付100%的有效载荷。我们的评估还表明,FDCommalm独立于应用域和FL计划所使用的神经网络结构。