Data-intensive workloads and applications, such as machine learning (ML), are fundamentally limited by traditional computing systems based on the von-Neumann architecture. As data movement operations and energy consumption become key bottlenecks in the design of computing systems, the interest in unconventional approaches such as Near-Data Processing (NDP), machine learning, and especially neural network (NN)-based accelerators has grown significantly. Emerging memory technologies, such as ReRAM and 3D-stacked, are promising for efficiently architecting NDP-based accelerators for NN due to their capabilities to work as both: High-density/low-energy storage and in/near-memory computation/search engine. In this paper, we present a survey of techniques for designing NDP architectures for NN. By classifying the techniques based on the memory technology employed, we underscore their similarities and differences. Finally, we discuss open challenges and future perspectives that need to be explored in order to improve and extend the adoption of NDP architectures for future computing platforms. This paper will be valuable for computer architects, chip designers and researchers in the area of machine learning.
翻译:数据密集的工作量和应用程序,如机器学习(ML),基本上受到基于von-Neumann结构的传统计算系统的限制。随着数据流动作业和能源消耗成为计算系统设计中的关键瓶颈,对非常规方法的兴趣,例如近数据处理、机器学习,特别是神经网络加速器等的兴趣已大增。ReRAM和3D堆叠式的记忆技术等新兴技术,由于它们既具有高密度/低能存储能力,又具有在/早期计算/搜索引擎中工作的能力,因此对为NNT高效设计基于NDP加速器的前景大有希望:高密度/低能量储存和在/早期计算/搜索引擎中消耗能源。在本文件中,我们对NNNN设计NDP结构的技术进行了调查。通过对所采用的记忆技术进行分类,我们强调其相似性和差异。最后,我们讨论了需要探讨的公开挑战和未来观点,以便改进和扩展NDP结构对未来计算平台的采用。这份文件对于计算机建筑师、芯片设计师和机器学习领域的研究人员将很有价值。