The difficulty of deploying various deep learning (DL) models on diverse DL hardwares has boosted the research and development of DL compilers in the community. Several DL compilers have been proposed from both industry and academia such as Tensorflow XLA and TVM. Similarly, the DL compilers take the DL models described in different DL frameworks as input, and then generate optimized codes for diverse DL hardwares as output. However, none of the existing survey has analyzed the unique design of the DL compilers comprehensively. In this paper, we perform a comprehensive survey of existing DL compilers by dissecting the commonly adopted design in details, with emphasis on the DL oriented multi-level IRs, and frontend/backend optimizations. Specifically, we provide a comprehensive comparison among existing DL compilers from various aspects. In addition, we present detailed analysis of the multi-level IR design and compiler optimization techniques. Finally, several insights are highlighted as the potential research directions of DL compiler. This is the first survey paper focusing on the unique design of DL compiler, which we hope can pave the road for future research towards the DL compiler.
翻译:在各种DL硬件上部署各种深层次学习(DL)模型的困难,促进了社区DL汇编者的研究和发展,从Tensorflow XLA和TVM等行业和学术界提出了若干DL汇编者,例如Tensorflow XLA和TVM。同样,DL汇编者将不同的DL框架描述的DL模型作为投入,然后为不同的DL硬件生成最佳代码作为输出。然而,现有的调查没有一项全面分析了DL汇编者的独特设计。在本文件中,我们对现有DL汇编者进行了全面调查,将共同采用的设计细分为细节,重点是DL导向的多层次IR,以及前端/后端优化。具体地说,我们从各方面对现有DL汇编者进行了全面比较。此外,我们提出了对多层次的IR设计和编译者优化技术的详细分析。最后,一些见解被强调为DL汇编者的潜在研究方向。这是第一份调查文件,重点是DL汇编者的独特设计,我们希望它能够为未来的研究铺路。