A label-efficient paradigm in computer vision is based on self-supervised contrastive pre-training on unlabeled data followed by fine-tuning with a small number of labels. Making practical use of a federated computing environment in the clinical domain and learning on medical images poses specific challenges. In this work, we propose FedMoCo, a robust federated contrastive learning (FCL) framework, which makes efficient use of decentralized unlabeled medical data. FedMoCo has two novel modules: metadata transfer, an inter-node statistical data augmentation module, and self-adaptive aggregation, an aggregation module based on representational similarity analysis. To the best of our knowledge, this is the first FCL work on medical images. Our experiments show that FedMoCo can consistently outperform FedAvg, a seminal federated learning framework, in extracting meaningful representations for downstream tasks. We further show that FedMoCo can substantially reduce the amount of labeled data required in a downstream task, such as COVID-19 detection, to achieve a reasonable performance.
翻译:计算机愿景中的标签效率范例是基于对无标签数据进行自我监督的对比性前培训,然后对少量标签进行微调。在临床领域实际使用联合计算环境并学习医疗图像,这带来了具体的挑战。在这项工作中,我们提议FedMoCo,一个强有力的联合对比性学习框架,使分散的无标签医疗数据得到有效利用。FedMoCo有两个新的模块:元数据传输、一个跨节统计数据增强模块,以及基于代表性相似性分析的集成模块。据我们所知,这是FCL关于医疗图像的首次工作。我们的实验表明,FDMoco在为下游任务获取有意义的代表性时,可以一贯地超越FedAvg这个有特色的联邦化学习框架。我们进一步表明,FedMoCo可以大量减少下游任务(如COVID-19检测)所需的标签数据数量,以达到合理的性能。