Reliable yet efficient evaluation of generalisation performance of a proposed architecture is crucial to the success of neural architecture search (NAS). Traditional approaches face a variety of limitations: training each architecture to completion is prohibitively expensive, early stopped validation accuracy may correlate poorly with fully trained performance, and model-based estimators require large training sets. We instead propose to estimate the final test performance based on a simple measure of training speed. Our estimator is theoretically motivated by the connection between generalisation and training speed, and is also inspired by the reformulation of a PAC-Bayes bound under the Bayesian setting. Our model-free estimator is simple, efficient, and cheap to implement, and does not require hyperparameter-tuning or surrogate training before deployment. We demonstrate on various NAS search spaces that our estimator consistently outperforms other alternatives in achieving better correlation with the true test performance rankings. We further show that our estimator can be easily incorporated into both query-based and one-shot NAS methods to improve the speed or quality of the search.
翻译:对拟议建筑的总体性能进行可靠而有效的评估,对于神经结构搜索的成功至关重要。 传统方法面临各种限制:对每个建筑进行完成培训的费用高得令人望而却步,早期停止的验证准确性可能与充分培训的性能不相干,而基于模型的估测员则需要大量的训练。我们提议根据简单的培训速度来估计最终测试性能。我们的估测员在理论上是受一般性能与培训速度之间联系的驱动,同时也受到重新制定受巴伊西亚环境约束的PAC-Bayes的启发。我们的无型天线测量仪简单、高效且廉价,不需要在部署前进行超离子仪调整或代孕培训。我们在各种NAS搜索空间上展示,我们的估测员在与真正的测试性能排级相比,始终优于其他选择。我们进一步表明,我们的估测员可以很容易地纳入基于查询的和一发式的NAS方法,以提高搜索速度或质量。