Federated Learning (FL) is a machine learning paradigm where many local nodes collaboratively train a central model while keeping the training data decentralized. This is particularly relevant for clinical applications since patient data are usually not allowed to be transferred out of medical facilities, leading to the need for FL. Existing FL methods typically share model parameters or employ co-distillation to address the issue of unbalanced data distribution. However, they also require numerous rounds of synchronized communication and, more importantly, suffer from a privacy leakage risk. We propose a privacy-preserving FL framework leveraging unlabeled public data for one-way offline knowledge distillation in this work. The central model is learned from local knowledge via ensemble attention distillation. Our technique uses decentralized and heterogeneous local data like existing FL approaches, but more importantly, it significantly reduces the risk of privacy leakage. We demonstrate that our method achieves very competitive performance with more robust privacy preservation based on extensive experiments on image classification, segmentation, and reconstruction tasks.
翻译:联邦学习(FL)是一种机器学习模式,许多地方节点在保持培训数据分散的同时合作培训一个中心模型,这与临床应用特别相关,因为病人数据通常不允许从医疗设施中转移,导致需要FL。现有的FL方法通常共享模型参数,或采用共同蒸馏方法来解决数据分布不平衡的问题。但是,它们也需要多轮同步通信,更重要的是,它们也存在隐私渗漏风险。我们提议一个隐私保护框架,利用未贴标签的公共数据进行单向离线知识蒸馏。中央模型是通过混合关注蒸馏从当地知识中学习的。我们的技术使用分散和多样化的当地数据,如现有的FL方法,但更重要的是,它大大降低了隐私渗漏的风险。我们证明,我们的方法在对图像分类、分解和重建任务进行广泛实验的基础上,以更强有力的隐私保护方式取得了非常有竞争力的业绩。