Segmentation is one of the most primary tasks in deep learning for medical imaging, owing to its multiple downstream clinical applications. However, generating manual annotations for medical images is time-consuming, requires high skill, and is an expensive effort, especially for 3D images. One potential solution is to aggregate knowledge from partially annotated datasets from multiple groups to collaboratively train global models using Federated Learning. To this end, we propose SegViz, a federated learning-based framework to train a segmentation model from distributed non-i.i.d datasets with partial annotations. The performance of SegViz was compared against training individual models separately on each dataset as well as centrally aggregating all the datasets in one place and training a single model. The SegViz framework using FedBN as the aggregation strategy demonstrated excellent performance on the external BTCV set with dice scores of 0.93, 0.83, 0.55, and 0.75 for segmentation of liver, spleen, pancreas, and kidneys, respectively, significantly ($p<0.05$) better (except spleen) than the dice scores of 0.87, 0.83, 0.42, and 0.48 for the baseline models. In contrast, the central aggregation model significantly ($p<0.05$) performed poorly on the test dataset with dice scores of 0.65, 0, 0.55, and 0.68. Our results demonstrate the potential of the SegViz framework to train multi-task models from distributed datasets with partial labels. All our implementations are open-source and available at https://anonymous.4open.science/r/SegViz-B746
翻译:由于医学成像的多种下游临床应用,深入学习医学成像的最主要任务之一是进行分解。然而,为医学成像制作人工说明是耗费时间的,需要高技能,而且是一项昂贵的工作,特别是3D图像。一个潜在的解决办法是,将来自多个群体部分附加说明的数据集的知识汇总起来,利用Freed Learning合作培训全球模型。为此,我们提议SegViz是一个基于学习的联邦框架,用于从分布的非i.i.i.d部分数据集中培训一个分解模型,并附有部分的公开说明。SegViz的性能与每个数据集单独培训的单个模型进行比较,同时集中汇集所有数据集,并培训单一模型。SegViz框架使用FedBN作为汇总战略,在外部BTCV集中展示了优异性业绩,其分数分别为0.93、0.83、0.55和0.75美元(Spealen)的分解模型,比0.87、0.42和0.48S(我们基本数据模型的0.65)的全基数和0.65、0.65(我们基本数据中的0.65)和0.65)测试模型显示。</s>