Pervasive computing promotes the installation of connected devices in our living spaces in order to provide services. Two major developments have gained significant momentum recently: an advanced use of edge resources and the integration of machine learning techniques for engineering applications. This evolution raises major challenges, in particular related to the appropriate distribution of computing elements along an edge-to-cloud continuum. About this, Federated Learning has been recently proposed for distributed model training in the edge. The principle of this approach is to aggregate models learned on distributed clients in order to obtain a new, more general model. The resulting model is then redistributed to clients for further training. To date, the most popular federated learning algorithm uses coordinate-wise averaging of the model parameters for aggregation. However, it has been shown that this method is not adapted in heterogeneous environments where data is not identically and independently distributed (non-iid). This corresponds directly to some pervasive computing scenarios where heterogeneity of devices and users challenges machine learning with the double objective of generalization and personalization. In this paper, we propose a novel aggregation algorithm, termed FedDist, which is able to modify its model architecture (here, deep neural network) by identifying dissimilarities between specific neurons amongst the clients. This permits to account for clients' specificity without impairing generalization. Furthermore, we define a complete method to evaluate federated learning in a realistic way taking generalization and personalization into account. Using this method, FedDist is extensively tested and compared with three state-of-the-art federated learning algorithms on the pervasive domain of Human Activity Recognition with smartphones.
翻译:渗透式计算促进在生活空间安装连接设备,以提供服务。最近出现了两大重大进展:先进使用边缘资源和整合机械学习技术用于工程应用。这一演变带来了重大挑战,特别是在边际至宽宽度连续体上适当分配计算元素方面。关于这一点,最近提议在边缘进行分布模式培训。这一方法的原则是将分布客户所学的模型汇总起来,以便获得一个新的、更普遍的模型。随后,将由此产生的模型重新分配给客户进一步培训。迄今为止,最受欢迎的联合学习算法使用协调的模型参数的平均值。然而,这一方法在数据分布不完全和独立(非二)的多种环境中并不适应。这直接与一些普遍的计算假设方案相对应,即设备和用户的异质性化和个性化双重目标对机器学习构成挑战。在这个文件中,我们提议一种新型的合并算法,即FedDdddist,它能够修改其模型结构(这里的、深度的智能算法,用来平均地计算模型结构,而不用精确性化的三种计算方法,用来对普通客户进行精确的计算。