Although recent scaling up approaches to training deep neural networks have proven to be effective, the computational intensity of large and complex models, as well as the availability of large-scale datasets, require deep learning frameworks to utilize scaling out techniques. Parallelization approaches and distribution requirements are not considered in the preliminary designs of most available distributed deep learning frameworks, and most of them still are not able to perform effective and efficient fine-grained inter-node communication. We present Phylanx that has the potential to alleviate these shortcomings. Phylanx offers a productivity-oriented frontend where user Python code is translated to a futurized execution tree that can be executed efficiently on multiple nodes using the C++ standard library for parallelism and concurrency (HPX), leveraging fine-grained threading and an active messaging task-based runtime system.
翻译:虽然最近扩大培训深层神经网络的方法已证明是有效的,但大型和复杂模型的计算强度以及大规模数据集的可用性需要深层次的学习框架来利用扩展技术。在大多数现有分布式深层学习框架的初步设计中,没有考虑到平行方法和分配要求,其中多数仍然无法进行有成效和高效率的细细细区分的跨节通信。我们介绍了有潜力减轻这些缺陷的Phylanx。Phyllanx提供了一个面向生产力的前端,用户Python代码被转换成一个未来化的执行树,可以利用C++标准图书馆的平行和同值图书馆(HPX)在多个节点上高效地执行,利用精细的线条和积极的传递任务运行时间系统。