In-Memory Computing (IMC) hardware using Memristive Crossbar Arrays (MCAs) are gaining popularity to accelerate Deep Neural Networks (DNNs) since it alleviates the "memory wall" problem associated with von-Neumann architecture. The hardware efficiency (energy, latency and area) as well as application accuracy (considering device and circuit non-idealities) of DNNs mapped to such hardware are co-dependent on network parameters, such as kernel size, depth etc. and hardware architecture parameters such as crossbar size. However, co-optimization of both network and hardware parameters presents a challenging search space comprising of different kernel sizes mapped to varying crossbar sizes. To that effect, we propose NAX -- an efficient neural architecture search engine that co-designs neural network and IMC based hardware architecture. NAX explores the aforementioned search space to determine kernel and corresponding crossbar sizes for each DNN layer to achieve optimal tradeoffs between hardware efficiency and application accuracy. Our results from NAX show that the networks have heterogeneous crossbar sizes across different network layers, and achieves optimal hardware efficiency and accuracy considering the non-idealities in crossbars. On CIFAR-10 and Tiny ImageNet, our models achieve 0.8%, 0.2% higher accuracy, and 17%, 4% lower EDAP (energy-delay-area product) compared to a baseline ResNet-20 and ResNet-18 models, respectively.
翻译:在模拟计算(IMC)中,使用Meristive Crossbar Arrays(MCAs)的模拟计算机硬件越来越受欢迎,以加速深神经网络(DNNS),因为它缓解了与 von-Neumann 建筑相关的“模拟墙”问题。在这种硬件上映的DNN的硬件效率(能源、延时和面积)以及应用精确度(考虑装置和电路非理想性)都取决于网络参数,如内核大小、深度等,以及跨标准尺寸等硬件结构参数。然而,网络和硬件参数的协同优化展示了具有挑战性的搜索空间,由不同内核大小的“模拟墙”问题组成。为此,我们建议NAX -- -- 一个高效的神经结构搜索引擎,共同设计神经网络网络网络网络网络网络和基于IMC硬件结构的网络。NAX探索上述搜索空间,以确定每个DNNN的内核和相应的跨标准尺寸大小,以达到硬件效率和应用精确度之间的最佳交易。我们从 NAX 网络的对比模型显示,跨网络的准确度是不同网络的,跨级和跨标准级的 REAR-REDER 和跨级。