While the application of differential privacy (DP) has been well-studied in cross-device federated learning (FL), there is a lack of work considering DP for cross-silo FL, a setting characterized by a limited number of clients each containing many data subjects. In cross-silo FL, usual notions of client-level privacy are less suitable as real-world privacy regulations typically concern in-silo data subjects rather than the silos themselves. In this work, we instead consider the more realistic notion of silo-specific item-level privacy, where silos set their own privacy targets for their local examples. Under this setting, we reconsider the roles of personalization in federated learning. In particular, we show that mean-regularized multi-task learning (MR-MTL), a simple personalization framework, is a strong baseline for cross-silo FL: under stronger privacy, silos are further incentivized to "federate" with each other to mitigate DP noise, resulting in consistent improvements relative to standard baseline methods. We provide a thorough empirical study of competing methods as well as a theoretical characterization of MR-MTL for a mean estimation problem, highlighting the interplay between privacy and cross-silo data heterogeneity. Our work serves to establish baselines for private cross-silo FL as well as identify key directions of future work in this area.
翻译:虽然不同隐私(DP)的应用在跨联合学习(FL)中得到了很好地研究,但对于跨筒仓FL(交叉筒仓FL)却缺乏考虑DP(DP)的工作,因为跨筒仓FL的特点是客户数量有限,每个包含许多数据主题。在跨筒仓FL,客户隐私的通常概念不太合适,因为现实世界隐私条例一般都涉及在筒仓内的数据主题,而不是筒仓本身。在这项工作中,我们反而认为,在筒仓内为自己的地方实例设定了自己的隐私目标,而采用单仓内特定物品隐私这一更现实的概念。在这种背景下,我们重新考虑个人化在交叠式学习中的作用。特别是,我们表明,一个简单的个性化的多任务学习(MR-MTL)是跨筒仓的坚实基准:在更紧密的隐私下,筒仓内进一步鼓励彼此“喂养”以缓解DP的噪音,从而与标准基线方法一致地改进。我们提出了关于相互竞争的方法的彻底的经验研究,作为MRML-MTL的理论性跨基领域,作为我们个人关系未来工作的重要方向的跨基的理论性分析。