Recent advances in distributed learning systems have introduced effective solutions for implementing collaborative artificial intelligence techniques in wireless communication networks. Federated learning approaches provide a model-aggregation mechanism among edge devices to achieve collaborative training, while ensuring data security, communication efficiency, and sharing computational overheads. On the other hand, limited transmission resources and complex communication environments remain significant bottlenecks to the efficient collaborations among edge devices, particularly within large-scale networks. To address such issues, this paper proposes an over-the-air (OTA) analog aggregation method designed for the distributed swarm learning (DSL), termed DSL-OTA, aiming to enhance communication efficiency, enable effective cooperation, and ensure privacy preserving. Incorporating multi-worker selection strategy with over-the-air aggregation not only makes the standard DSL based on single best worker contributing to global model update to become more federated, but also secures the aggregation from potential risks of data leakage. Our theoretical analyses verify the advantages of the proposed DSL-OTA algorithm in terms of fast convergence rate and low communication costs. Simulation results reveal that our DSL-OTA outperforms the other existing methods by achieving better learning performance under both homogeneous and heterogeneous dataset settings.
翻译:分布式学习系统的最新进展为在无线通信网络中实现协作人工智能技术提供了有效解决方案。联邦学习方法通过边缘设备间的模型聚合机制实现协同训练,同时确保数据安全、通信效率并分摊计算开销。然而,有限的传输资源和复杂的通信环境仍然是边缘设备间高效协作的重要瓶颈,尤其在大规模网络中更为突出。为解决这些问题,本文提出一种面向分布式群体学习(DSL)设计的空中(OTA)模拟聚合方法,称为DSL-OTA,旨在提升通信效率、实现有效协作并确保隐私保护。将多工作者选择策略与空中聚合相结合,不仅使基于单一最优工作者贡献的经典DSL全局模型更新机制更具联邦特性,还能防范数据泄露的潜在风险。理论分析验证了所提DSL-OTA算法在快速收敛速率和低通信成本方面的优势。仿真结果表明,在均匀和非均匀数据集设置下,DSL-OTA均能获得更优的学习性能,优于现有其他方法。