In the past few decades, machine learning has revolutionized data processing for large scale applications. Simultaneously, increasing privacy threats in trending applications led to the redesign of classical data training models. In particular, classical machine learning involves centralized data training, where the data is gathered, and the entire training process executes at the central server. Despite significant convergence, this training involves several privacy threats on participants' data when shared with the central cloud server. To this end, federated learning has achieved significant importance over distributed data training. In particular, the federated learning allows participants to collaboratively train the local models on local data without revealing their sensitive information to the central cloud server. In this paper, we perform a convergence comparison between classical machine learning and federated learning on two publicly available datasets, namely, logistic-regression-MNIST dataset and image-classification-CIFAR-10 dataset. The simulation results demonstrate that federated learning achieves higher convergence within limited communication rounds while maintaining participants' anonymity. We hope that this research will show the benefits and help federated learning to be implemented widely.
翻译:在过去几十年中,机器学习使大规模应用的数据处理发生了革命性的变化。与此同时,趋势应用中的隐私威胁日益增加,导致古典数据培训模式的重新设计。特别是,古典机器学习涉及中央数据培训,在收集数据的地方,整个培训过程在中央服务器进行。尽管这种培训有相当的趋同性,但在与中央云服务器共享时,参与者的数据受到若干隐私威胁。为此,联合学习在分发数据培训中取得了重要的意义。特别是,联合学习使参与者能够合作培训当地数据模型,而不会向中央云服务器透露其敏感信息。在本文中,我们比较了经典机器学习和在两个公开数据集(即后勤-反向-MNIST数据集和图像分类-CIFAR-10数据集)上进行的联合学习。模拟结果显示,在有限的通信回合中,热学在保持参与者的匿名性的同时,实现了更高程度的趋同。我们希望,这项研究将展示其益处,并有助于在广泛实施进化学习。