Advanced adversarial attacks such as membership inference and model memorization can make federated learning (FL) vulnerable and potentially leak sensitive private data. Local differentially private (LDP) approaches are gaining more popularity due to stronger privacy notions and native support for data distribution compared to other differentially private (DP) solutions. However, DP approaches assume that the FL server (that aggregates the models) is honest (run the FL protocol honestly) or semi-honest (run the FL protocol honestly while also trying to learn as much information as possible). These assumptions make such approaches unrealistic and unreliable for real-world settings. Besides, in real-world industrial environments (e.g., healthcare), the distributed entities (e.g., hospitals) are already composed of locally running machine learning models (this setting is also referred to as the cross-silo setting). Existing approaches do not provide a scalable mechanism for privacy-preserving FL to be utilized under such settings, potentially with untrusted parties. This paper proposes a new local differentially private FL (named LDPFL) protocol for industrial settings. LDPFL can run in industrial settings with untrusted entities while enforcing stronger privacy guarantees than existing approaches. LDPFL shows high FL model performance (up to 98%) under small privacy budgets (e.g., epsilon = 0.5) in comparison to existing methods.
翻译:高级对抗性攻击,如会籍推断和模型记忆化等,可以使联谊学习(FL)容易受害,并有可能泄漏敏感的私人数据。由于隐私概念的加强和当地对数据分配的支持,与其他差别私人(DP)解决方案相比,当地有差别的私人(LDP)方法越来越受欢迎。然而,DP方法假定FL服务器(综合模型)是诚实的(诚实地运行FL协议)或半诚实的(诚实地运行FL协议,同时努力尽可能多地了解信息 ) 。这些假设使得这种方法在现实世界环境中不现实世界环境中不切实际和不可靠。此外,在现实世界的工业环境中(例如保健),分布的实体(例如医院)已经由本地运行的机器学习模式组成(这个环境也被称为交叉筒式设置 ) 。 现有的办法并不能提供一种可缩放机制,用于在这种环境下使用隐私保护FLL(称为LDP) 协议的新的本地有差别的私人(LLFL) 。LFL(在现有的保密性预算下,可以进行更强的比较,而没有信任的LFLFL) 预算下的高级的运行方式。