The general approach taken when training deep learning classifiers is to save the parameters after every few iterations, train until either a human observer or a simple metric-based heuristic decides the network isn't learning anymore, and then backtrack and pick the saved parameters with the best validation accuracy. Simple methods are used to determine if a neural network isn't learning anymore because, as long as it's well after the optimal values are found, the condition doesn't impact the final accuracy of the model. However from a runtime perspective, this is of great significance to the many cases where numerous neural networks are trained simultaneously (e.g. hyper-parameter tuning). Motivated by this, we introduce a statistical significance test to determine if a neural network has stopped learning. This stopping criterion appears to represent a happy medium compared to other popular stopping criterions, achieving comparable accuracy to the criterions that achieve the highest final accuracies in 77% or fewer epochs, while the criterions which stop sooner do so with an appreciable loss to final accuracy. Additionally, we use this as the basis of a new learning rate scheduler, removing the need to manually choose learning rate schedules and acting as a quasi-line search, achieving superior or comparable empirical performance to existing methods.
翻译:在培训深层次学习分类师时,通常的做法是在每次迭代后保存参数,培训直到人类观察者或简单的光学理论决定网络不再学习,然后回溯和以最准确的校验精度选择节省的参数。使用简单的方法来确定神经网络是否不再学习,因为只要在找到最佳值之后很久,该条件并不影响模型的最终准确性。然而,从运行时间的角度来看,这对于许多神经网络同时接受培训的许多案例(例如超参数调)非常重要。为此,我们引入了具有统计意义的测试,以确定神经网络是否停止学习。这个停止标准似乎代表一种快乐的介质,而与其他流行的停止标准相比,它达到了77%或更少的达到最高最终理解度的标准具有可比性,而停止的标准则更早于最终准确性。此外,我们用这个标准作为新的学习率表的基础,消除了现有可比较的手动学习率,从而选择了一种可比较的手动学习率。