With the increased legislation around data privacy, federated learning (FL) has emerged as a promising technique that allows the clients (end-user) to collaboratively train deep learning (DL) models without transferring and storing the data in a centralized, third-party server. Despite the theoretical success, FL is yet to be adopted in real-world systems due to the hardware, computing, and various infrastructure constraints presented by the edge and mobile devices of the clients. As a result, simulated datasets, models, and experiments are heavily used by the FL research community to validate their theories and findings. We introduce TorchFL, a performant library for (i) bootstrapping the FL experiments, (ii) executing them using various hardware accelerators, (iii) profiling the performance, and (iv) logging the overall and agent-specific results on the go. Being built on a bottom-up design using PyTorch and Lightning, TorchFL provides ready-to-use abstractions for models, datasets, and FL algorithms, while allowing the developers to customize them as and when required.
翻译:随着关于数据隐私的立法的增加,联合学习(FL)已成为一种有希望的技术,使客户(最终用户)能够合作培训深层次学习模式,而无需在集中的第三方服务器上转让和储存数据。尽管在理论上取得了成功,但由于客户的边缘和移动设备带来的硬件、计算和各种基础设施限制,FL尚未在现实世界系统中采用FL。结果,FL研究界大量使用模拟数据集、模型和实验来验证其理论和发现。我们引入了TorchFL,这是一个用于(一) 制备FL实验的表演图书馆,(二) 使用各种硬件加速器执行这些模型,(三) 描述性能,(四) 记录在移动上的总体和特定代理结果。TorchFL在使用PyTorrch和Lightning的底层设计的基础上,TorchFL为模型、数据集和FL算法提供了现用的抽象数据,同时允许开发商按需要和按需要定制这些模型、数据集和FL算法。