Neural Architecture Search (NAS) has recently become a topic of great interest. However, there is a potentially impactful issue within NAS that remains largely unrecognized: noise. Due to stochastic factors in neural network initialization, training, and the chosen train/validation dataset split, the performance evaluation of a neural network architecture, which is often based on a single learning run, is also stochastic. This may have a particularly large impact if a dataset is small. We therefore propose to reduce the noise by having architecture evaluations comprise averaging of scores over multiple network training runs using different random seeds and cross-validation. We perform experiments for a combinatorial optimization formulation of NAS in which we vary noise reduction levels. We use the same computational budget for each noise level in terms of network training runs, i.e., we allow less architecture evaluations when averaging over more training runs. Multiple search algorithms are considered, including evolutionary algorithms which generally perform well for NAS. We use two publicly available datasets from the medical image segmentation domain where datasets are often limited and variability among samples is often high. Our results show that reducing noise in architecture evaluations enables finding better architectures by all considered search algorithms.
翻译:最近,神经结构搜索(NAS)已成为一个引起极大兴趣的话题。然而,NAS内部存在一个潜在的影响性问题,但基本上未被承认:噪音。由于神经网络初始化、培训和选定的火车/校验数据集分离的随机因素,神经网络结构的性能评估通常以单一的学习运行为基础,也是随机的。如果数据集小,这可能会产生特别大的影响。因此,我们提议通过结构评估来减少噪音,其中包括使用不同随机种子和交叉校验的多种网络培训运行的平均分数。我们实验了NAS的组合式优化配方,在这种配方中我们减少噪音的程度不同。我们在网络培训运行方面使用相同的计算预算,也就是说,在平均进行更多的培训运行时,我们允许较少地进行结构评估。我们考虑了多种搜索算法,包括通常对NAS表现良好的进化算法。我们使用两种公开的数据集,从医学图像分割域中得出,那里的数据集往往有限,而且样本之间变化性也比较大。我们的算法结果显示,通过搜索结构来降低噪音。