Practical use of neural networks often involves requirements on latency, energy and memory among others. A popular approach to find networks under such requirements is through constrained Neural Architecture Search (NAS). However, previous methods use complicated predictors for the accuracy of the network. Those predictors are hard to interpret and sensitive to many hyperparameters to be tuned, hence, the resulting accuracy of the generated models is often harmed. In this work we resolve this by introducing Bilinear Interpretable Neural Architecture Search (BINAS), that is based on an accurate and simple bilinear formulation of both an accuracy estimator and the expected resource requirement, together with a scalable search method with theoretical guarantees. The simplicity of our proposed estimator together with the intuitive way it is constructed bring interpretability through many insights about the contribution of different design choices. For example, we find that in the examined search space, adding depth and width is more effective at deeper stages of the network and at the beginning of each resolution stage. Our experiments show that BINAS generates comparable to or better architectures than other state-of-the-art NAS methods within a reduced marginal search cost, while strictly satisfying the resource constraints.
翻译:对神经网络的实际使用往往涉及对延缓、能量和记忆等的要求。在这类要求下寻找网络的流行方法是限制神经结构搜索(NAS)。然而,以前的方法是使用复杂的预测器来测量网络的准确性。这些预测器很难解释,而且对许多超参数难以调整,因此所产生的模型的准确性往往受到损害。在这项工作中,我们采用双线可解释神经结构搜索(BINAS)来解决这个问题,这种搜索是以精确估计器和预期资源要求的精确和简单的双线性配方为基础,同时采用有理论保证的可缩放搜索方法。我们提议的估测器的简单性以及其构建的直观方法,通过对不同设计选择的贡献的许多洞察力,带来了解释性。例如,我们发现在所审查的搜索空间中,增加深度和宽度在网络的更深阶段和每个分辨率阶段都更为有效。我们的实验显示,BINASS在严格满足资源边际成本的同时,在严格满足资源边际搜索限制的范围内,其结构可以比其他状态或更好。