Test Input Prioritizers (TIP) for Deep Neural Networks (DNN) are an important technique to handle the typically very large test datasets efficiently, saving computation and labeling costs. This is particularly true for large-scale, deployed systems, where inputs observed in production are recorded to serve as potential test or training data for the next versions of the system. Feng et. al. propose DeepGini, a very fast and simple TIP, and show that it outperforms more elaborate techniques such as neuron- and surprise coverage. In a large-scale study (4 case studies, 8 test datasets, 32'200 trained models) we verify their findings. However, we also find that other comparable or even simpler baselines from the field of uncertainty quantification, such as the predicted softmax likelihood or the entropy of the predicted softmax likelihoods perform equally well as DeepGini.
翻译:深神经网络(DNN)的测试输入优先排序器(TIP)是高效处理通常非常庞大的测试数据集、节省计算和标签成本的重要技术。对于大规模部署的系统来说尤其如此,因为生产过程中观察到的投入被记录为系统下一版本的潜在测试或培训数据。 Feng 等人建议采用非常快速和简单的 DeepGini(DeepGini),这是一个非常快速和简单的TIP,并表明它比神经和突袭覆盖等更为复杂的技术要好。在一项大规模研究中(4个案例研究,8个测试数据集,32'200个经过培训的模型),我们核实了它们的调查结果。然而,我们还发现,从不确定性量化领域看,其他可比的甚至更简单的基线,例如预测的软通气可能性或预测的软通气可能性的英特罗普,同样也与DeepGini(DeepGini)一样有效。