Neural Architecture Search (NAS) has recently become a topic of great interest. However, there is a potentially impactful issue within NAS that remains largely unrecognized: noise. Due to stochastic factors in neural network initialization, training, and the chosen train/validation dataset split, the performance evaluation of a neural network architecture, which is often based on a single learning run, is also stochastic. This may have a particularly large impact if a dataset is small. We therefore propose to reduce this noise by evaluating architectures based on average performance over multiple network training runs using different random seeds and cross-validation. We perform experiments for a combinatorial optimization formulation of NAS in which we vary noise reduction levels. We use the same computational budget for each noise level in terms of network training runs, i.e., we allow less architecture evaluations when averaging over more training runs. Multiple search algorithms are considered, including evolutionary algorithms which generally perform well for NAS. We use two publicly available datasets from the medical image segmentation domain where datasets are often limited and variability among samples is often high. Our results show that reducing noise in architecture evaluations enables finding better architectures by all considered search algorithms.
翻译:最近,神经结构搜索(NAS)已成为一个引起极大兴趣的话题。然而,NAS内部存在一个潜在的影响性问题,但基本上未被承认:噪音。由于神经网络初始化、培训和选定的火车/校验数据集分离的随机因素,神经网络结构的性能评估通常以单一的学习运行为基础,也是随机性的。如果数据集小,这可能会产生特别大的影响。因此,我们提议通过评估基于使用不同随机种子和交叉校验的多种网络培训运行的平均性能的建筑来减少这种噪音。我们实验了NAS的组合式优化配制,在这种配制中,我们减少噪音的程度各不相同。我们使用相同的计算预算来计算网络培训运行中的每种噪音,也就是说,在平均超过培训运行时,我们允许较少的结构评估。我们考虑了多种搜索算法,包括通常对NAS表现良好的进化算法。我们使用两种公开的数据集。我们使用两种公开的数据集,在医学图像分割域域中,那里的数据集往往有限,而且样本之间的可变性也很高。我们的算法结果显示通过搜索结构来降低噪音。