The traditional approach to distributed machine learning is to adapt learning algorithms to the network, e.g., reducing updates to curb overhead. Networks based on intelligent edge, instead, make it possible to follow the opposite approach, i.e., to define the logical network topology em around the learning task to perform, so as to meet the desired learning performance. In this paper, we propose a system model that captures such aspects in the context of supervised machine learning, accounting for both learning nodes (that perform computations) and information nodes (that provide data). We then formulate the problem of selecting (i) which learning and information nodes should cooperate to complete the learning task, and (ii) the number of iterations to perform, in order to minimize the learning cost while meeting the target prediction error and execution time. After proving important properties of the above problem, we devise an algorithm, named DoubleClimb, that can find a 1+1/|I|-competitive solution (with I being the set of information nodes), with cubic worst-case complexity. Our performance evaluation, leveraging a real-world network topology and considering both classification and regression tasks, also shows that DoubleClimb closely matches the optimum, outperforming state-of-the-art alternatives.
翻译:分布式机器学习的传统方法是将学习算法适应网络,例如,减少更新,以遏制间接费用; 以智能边缘为基础的网络,相反地,可以采取相反的方法,即界定学习任务周围的逻辑网络地形学,以便达到预期的学习业绩; 在本文中,我们提出一个系统模型,在监督的机器学习的背景下捕捉这些方面,既考虑到学习节点(进行计算),又考虑到信息节点(提供数据); 然后我们提出选择问题:(一) 哪些学习和信息节点应当合作完成学习任务,以及(二) 要执行的迭代数量,以便在达到目标预测错误和执行时间的同时尽量减少学习成本。在证明了上述问题的重要特性之后,我们设计了一个名为“双立”的算法,可以找到一个1+1/I ⁇ -竞争性的解决方案(我就是一套信息节点),而这个方法的复杂度是每立方子。我们的业绩评估,利用一个实际网络的顶级和考虑最佳的分类和回归任务。