With the development of federated learning (FL), mobile devices (MDs) are able to train their local models with private data and sends them to a central server for aggregation, thereby preventing sensitive raw data leakage. In this paper, we aim to improve the training performance of FL systems in the context of wireless channels and stochastic energy arrivals of MDs. To this purpose, we dynamically optimize MDs' transmission power and training task scheduling. We first model this dynamic programming problem as a constrained Markov decision process (CMDP). Due to high dimensions rooted from our CMDP problem, we propose online stochastic learning methods to simplify the CMDP and design online algorithms to obtain an efficient policy for all MDs. Since there are long-term constraints in our CMDP, we utilize Lagrange multipliers approach to tackle this issue. Furthermore, we prove the convergence of the proposed online stochastic learning algorithm. Numerical results indicate that the proposed algorithms can achieve better performance than the benchmark algorithms.
翻译:随着联合学习(FL)的发展,移动设备(MDs)能够用私人数据对本地模型进行培训,并将其发送到中央服务器,以便汇总,从而防止敏感的原始数据泄漏。在本文件中,我们的目标是提高FL系统在无线频道和MDs抵达的随机能源方面的培训绩效。为此,我们动态地优化MDs的传输动力和培训任务时间安排。我们首先将这种动态编程问题作为有限的Markov决定程序(CMDP)来模型。由于我们CMDP问题产生的高维度,我们建议采用在线随机化学习方法来简化CMDP,并设计在线算法,以便为所有MDs制定有效的政策。由于我们的CMDP存在长期限制,我们使用Lagrange乘法来解决这个问题。此外,我们证明了拟议的在线Stochacistic学习算法的趋同性。数字结果表明,拟议的算法可以比基准算法取得更好的性能。