This paper proposes FedPOD (Proportionally Orchestrated Derivative) for optimizing learning efficiency and communication cost in federated learning among multiple clients. Inspired by FedPIDAvg, we define a round-wise task for FedPOD to enhance training efficiency. FedPIDAvg achieved performance improvement by incorporating the training loss reduction for prediction entropy as weights using differential terms. Furthermore, by modeling data distribution with a Poisson distribution and using a PID controller, it reduced communication costs even in skewed data distribution. However, excluding participants classified as outliers based on the Poisson distribution can limit data utilization. Additionally, PID controller requires the same participants to be maintained throughout the federated learning process as it uses previous rounds' learning information in the current round. In our approach, FedPOD addresses these issues by including participants excluded as outliers, eliminating dependency on previous rounds' learning information, and applying a method for calculating validation loss at each round. In this challenge, FedPOD presents comparable performance to FedPIDAvg in metrics of Dice score, 0.78, 0.71 and 0.72 for WT, ET and TC in average, and projected convergence score, 0.74 in average. Furthermore, the concept of FedPOD draws inspiration from Kubernetes' smallest computing unit, POD, designed to be compatible with Kubernetes auto-scaling. Extending round-wise tasks of FedPOD to POD units allows flexible design by applying scale-out similar to Kubernetes' auto-scaling. This work demonstrated the potentials of FedPOD to enhance federated learning by improving efficiency, flexibility, and performance in metrics.
翻译:本文提出FedPOD(比例编排导数)方法,用于优化多客户端联邦学习中的学习效率与通信成本。受FedPIDAvg启发,我们为FedPOD定义了逐轮任务以提升训练效率。FedPIDAvg通过将训练损失对预测熵的减少量作为权重并结合微分项,实现了性能提升。此外,通过使用泊松分布建模数据分布并结合PID控制器,该方法即使在偏斜数据分布下也能降低通信成本。然而,基于泊松分布排除被归类为异常值的参与者会限制数据利用率。另外,PID控制器需要在联邦学习全程保持相同参与者集合,因其在当前轮次使用了历史轮次的学习信息。在本方法中,FedPOD通过以下方式解决这些问题:纳入被排除的异常值参与者、消除对历史轮次学习信息的依赖,并采用逐轮验证损失计算方法。在本次挑战中,FedPOD在Dice分数指标上表现出与FedPIDAvg相当的性能,WT、ET和TC区域的平均值分别为0.78、0.71和0.72,平均收敛分数为0.74。此外,FedPOD的设计理念借鉴了Kubernetes的最小计算单元POD,旨在兼容Kubernetes自动扩缩容机制。将FedPOD的逐轮任务扩展至POD单元,可通过类似Kubernetes自动扩缩容的横向扩展方式实现灵活设计。本工作证明了FedPOD在提升联邦学习效率、灵活性和性能指标方面的潜力。