The widespread adoption of Deep Neural Networks (DNNs) in important domains raises questions about the trustworthiness of DNN outputs. Even a highly accurate DNN will make mistakes some of the time, and in settings like self-driving vehicles these mistakes must be quickly detected and properly dealt with in deployment. Just as our community has developed effective techniques and mechanisms to monitor and check programmed components, we believe it is now necessary to do the same for DNNs. In this paper we present DNN self-checking as a process by which internal DNN layer features are used to check DNN predictions. We detail SelfChecker, a self-checking system that monitors DNN outputs and triggers an alarm if the internal layer features of the model are inconsistent with the final prediction. SelfChecker also provides advice in the form of an alternative prediction. We evaluated SelfChecker on four popular image datasets and three DNN models and found that SelfChecker triggers correct alarms on 60.56% of wrong DNN predictions, and false alarms on 2.04% of correct DNN predictions. This is a substantial improvement over prior work (SELFORACLE, DISSECTOR, and ConfidNet). In experiments with self-driving car scenarios, SelfChecker triggers more correct alarms than SELFORACLE for two DNN models (DAVE-2 and Chauffeur) with comparable false alarms. Our implementation is available as open source.
翻译:在重要领域广泛采用深神经网络(DNN)的做法使人对DNN产出的可信度产生疑问。即使高度准确的DNN将犯一些错误,在诸如自行驾驶的车辆等环境下,也必须迅速发现和适当处理这些错误。正如我们的社区已经开发了有效的技术和机制来监测和检查编程组件一样,我们认为现在有必要对DNN采取同样的做法。在本文件中,我们提出DNN的自我检查作为使用DNN内部的功能来检查DN预测的过程。我们详细介绍了SefCer,这是监测DNN产出的自我检查系统,如果该模型的内部层特征与最后预测不一致,则触发警报。自我检查者还以替代预测的形式提供了建议。我们对4个受欢迎的图像数据集和3个DNNNM模型进行了自我检查,发现“自我检查”触发了60.56%的DNNN的公开预测的警报,以及2.04 %的DNNN预测来源的错误警报。这是对先前工作的大幅改进(SEFOR,SDARC 和SLAF ARC ) 的自我警报模型比SUDADAD Adrifrifor AdroD Arrisal Adro) 更适合。