For safety assurance of deep neural networks (DNNs), out-of-distribution (OoD) monitoring techniques are essential as they filter spurious input that is distant from the training dataset. This paper studies the problem of systematically testing OoD monitors to avoid cases where an input data point is tested as in-distribution by the monitor, but the DNN produces spurious output predictions. We consider the definition of "in-distribution" characterized in the feature space by a union of hyperrectangles learned from the training dataset. Thus the testing is reduced to finding corners in hyperrectangles distant from the available training data in the feature space. Concretely, we encode the abstract location of every data point as a finite-length binary string, and the union of all binary strings is stored compactly using binary decision diagrams (BDDs). We demonstrate how to use BDDs to symbolically extract corners distant from all data points within the training set. Apart from test case generation, we explain how to use the proposed corners to fine-tune the DNN to ensure that it does not predict overly confidently. The result is evaluated over examples such as number and traffic sign recognition.
翻译:对于深神经网络(DNN)的安全保障而言,分配外(OoD)监测技术至关重要,因为它们过滤了远离培训数据集的虚假输入。本文研究系统测试OOD显示器的问题,以避免出现由显示器在分布时测试输入数据点的情况,但DNN产生虚假的输出预测。我们认为,从培训数据集中学习到的超矩形联盟在特征空间中给“在分配中”所定性的定义。因此,测试减少了,以在远离功能空间现有培训数据的超矩形中查找角。具体地说,我们把每个数据点的抽象位置编码成一个固定的二进制字符串,而所有二进字符串的组合都是使用二进制决定图(BDDs)集中储存的。我们演示了如何使用BDDS来象征性地提取远离培训数据集中所有数据点的角落。除了测试案例生成之外,我们解释如何使用拟议的角对DNN进行微调,以确保它不会过于自信地预测流量。我们用这些例子来评估结果。