In-memory computing (IMC) on a monolithic chip for deep learning faces dramatic challenges on area, yield, and on-chip interconnection cost due to the ever-increasing model sizes. 2.5D integration or chiplet-based architectures interconnect multiple small chips (i.e., chiplets) to form a large computing system, presenting a feasible solution beyond a monolithic IMC architecture to accelerate large deep learning models. This paper presents a new benchmarking simulator, SIAM, to evaluate the performance of chiplet-based IMC architectures and explore the potential of such a paradigm shift in IMC architecture design. SIAM integrates device, circuit, architecture, network-on-chip (NoC), network-on-package (NoP), and DRAM access models to realize an end-to-end system. SIAM is scalable in its support of a wide range of deep neural networks (DNNs), customizable to various network structures and configurations, and capable of efficient design space exploration. We demonstrate the flexibility, scalability, and simulation speed of SIAM by benchmarking different state-of-the-art DNNs with CIFAR-10, CIFAR-100, and ImageNet datasets. We further calibrate the simulation results with a published silicon result, SIMBA. The chiplet-based IMC architecture obtained through SIAM shows 130$\times$ and 72$\times$ improvement in energy-efficiency for ResNet-50 on the ImageNet dataset compared to Nvidia V100 and T4 GPUs.
翻译:用于深层学习的单一芯片上的模拟(IMC)在地区、产量和芯片互连费用方面面临巨大的挑战,因为模型规模不断增加。 2.5D整合或基于芯片的架构将多个小芯片(即芯片)连接成一个大型的计算系统,提出了超越单一的IMC架构的可行解决方案,以加速大型深层学习模型。本文件介绍了一个新的基准模拟器,SIAM,以评价基于芯片的IMC架构的性能,并探讨IM结构设计中这种范式转变的潜力。SIM网络集成设备、电路、结构、网络对芯片的连接、网络对子的连接以及DRAM接入模型,以形成一个大型的、超越单一的IM结构,以加速大型的深层神经网络(DNNIS),适应各种基于网络的结构和配置,并能够对IM结构设计进行高效的探索。我们通过对IM网络设备、电网、网络、网络和网络上100美元的模式进行灵活度和模拟,通过不同州-10级的IM模型显示我们所出版的S-RO的S-ROAS-S-S-S-S-ROA的S-S-S-S-S-S-S-S-S-SIM-S-S-S-SIM-SIM-S-S-S-SLAS-S-S-S-S-S-S-S-S-S-S-S-SAS-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SLIM-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SL-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-