With the wide and deep adoption of deep learning models in real applications, there is an increasing need to model and learn the representations of the neural networks themselves. These models can be used to estimate attributes of different neural network architectures such as the accuracy and latency, without running the actual training or inference tasks. In this paper, we propose a neural architecture representation model that can be used to estimate these attributes holistically. Specifically, we first propose a simple and effective tokenizer to encode both the operation and topology information of a neural network into a single sequence. Then, we design a multi-stage fusion transformer to build a compact vector representation from the converted sequence. For efficient model training, we further propose an information flow consistency augmentation and correspondingly design an architecture consistency loss, which brings more benefits with less augmentation samples compared with previous random augmentation strategies. Experiment results on NAS-Bench-101, NAS-Bench-201, DARTS search space and NNLQP show that our proposed framework can be used to predict the aforementioned latency and accuracy attributes of both cell architectures and whole deep neural networks, and achieves promising performance.
翻译:随着在实际应用中广泛和深入采用深层学习模型,越来越需要建模和学习神经网络本身的表达方式。这些模型可用于估算不同神经网络结构的属性,例如精确度和延迟度,而无需实际培训或推断任务。在本文中,我们提出了一个神经结构代表模式,可用于整体地估计这些属性。具体地说,我们首先提议一个简单有效的代号,将神经网络的运行和地形信息编码为单一序列。然后,我们设计一个多阶段聚变变变变器,从转换序列中构建一个紧凑的矢量代表方式。为了进行高效的模型培训,我们进一步建议信息流动一致性增强,并相应设计一个结构一致性损失,与以往随机增强战略相比,增加量较少的样本带来更多好处。NAS-Bench-101、NAS-Bench-201、DARSS搜索空间和NNLQP的实验结果显示,我们提议的框架可以用来预测细胞结构和整个深神经网络的上述定位和精确度和精确度特征,并实现有希望的业绩。