Realistic use of neural networks often requires adhering to multiple constraints on latency, energy and memory among others. A popular approach to find fitting networks is through constrained Neural Architecture Search (NAS). However, previous methods use complicated predictors for the accuracy of the network. Those predictors are hard to interpret and sensitive to many hyperparameters to be tuned, hence, the resulting accuracy of the generated models is often harmed. In this work we resolve this by introducing Interpretable Integer Quadratic programming Neural Architecture Search (IQNAS), that is based on an accurate and simple quadratic formulation of both the accuracy predictor and the expected resource requirement, together with a scalable search method with theoretical guarantees. The simplicity of our proposed predictor together with the intuitive way it is constructed bring interpretability through many insights about the contribution of different design choices. For example, we find that in the examined search space, adding depth and width is more effective at deeper stages of the network and at the beginning of each resolution stage. Our experiments show that IQNAS generates comparable to or better architectures than other state-of-the-art NAS methods within a reduced search cost for each additional generated network, while strictly satisfying the resource constraints.
翻译:对神经网络的现实使用往往要求坚持对延缓度、能量和记忆等多种限制。一个寻找安装网络的流行方法是通过限制神经结构搜索(NAS) 。然而,以前的方法对网络的准确性使用复杂的预测器。这些预测器很难解释,而且对许多超参数难以调整,因此,所产生的模型的准确性往往受到损害。在这项工作中,我们通过引入可解释的 Integer 量度编程神经结构搜索(IQNAS)来解决这个问题,该方法基于精确预测器和预期资源要求的精确和简单的四边配方配方,同时采用具有理论保证的可缩放搜索方法。我们提议的预测器的简单性与它所构建的直观性方法一道,通过对不同设计选择的贡献的许多洞察来带来解释性。例如,我们发现在所研究的搜索空间中,增加深度和宽度在网络的更深阶段和每个分辨率阶段都更为有效。我们的实验显示,在严格地确定网络内产生的额外搜索成本的同时,IQNAS在每一种状态下产生比其他的搜索方法可比或更好的结构。