Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have increased significantly. Consequently, it has become important to pay attention to these footprint metrics of a model as well, not just its quality. We present and motivate the problem of efficiency in deep learning, followed by a thorough survey of the five core areas of model efficiency (spanning modeling techniques, infrastructure, and hardware) and the seminal work there. We also present an experiment-based guide along with code, for practitioners to optimize their model training and deployment. We believe this is the first comprehensive survey in the efficient deep learning space that covers the landscape of model efficiency from modeling techniques to hardware support. Our hope is that this survey would provide the reader with the mental model and the necessary understanding of the field to apply generic efficiency techniques to immediately get significant improvements, and also equip them with ideas for further research and experimentation to achieve additional gains.
翻译:深层学习使计算机视野、自然语言理解、语音识别、信息检索等领域发生了革命性的变化,然而,随着深层学习模式的逐步改进,其参数、长期性、培训所需资源等数量都大大增加,因此,必须注意模型的足迹度量,而不仅仅是其质量。我们提出并激发了深层学习的效率问题,随后对模型效率的五个核心领域(包括建模技术、基础设施和硬件)和其中的开创性工作进行了彻底调查。我们还提出了一个实验性指南和代码,供从业人员优化其示范培训和部署。我们认为这是对高效深层学习空间的第一次全面调查,涵盖模型从建模技术到硬件支持的效率景观。我们希望这一调查将为读者提供心理模型和对实地的必要理解,以便运用通用效率技术立即获得重大改进,并为他们提供进一步研究和实验的构想,以取得更多成果。