The use of deep neural networks (DNNs) in safety-critical applications like mobile health and autonomous driving is challenging due to numerous model-inherent shortcomings. These shortcomings are diverse and range from a lack of generalization over insufficient interpretability to problems with malicious inputs. Cyber-physical systems employing DNNs are therefore likely to suffer from safety concerns. In recent years, a zoo of state-of-the-art techniques aiming to address these safety concerns has emerged. This work provides a structured and broad overview of them. We first identify categories of insufficiencies to then describe research activities aiming at their detection, quantification, or mitigation. Our paper addresses both machine learning experts and safety engineers: The former ones might profit from the broad range of machine learning topics covered and discussions on limitations of recent methods. The latter ones might gain insights into the specifics of modern ML methods. We moreover hope that our contribution fuels discussions on desiderata for ML systems and strategies on how to propel existing approaches accordingly.
翻译:在诸如移动健康和自主驾驶等安全关键应用中使用深神经网络(DNNs)具有挑战性,原因是许多模型内在缺陷,这些缺陷多种多样,从缺乏对可解释性不足的概括性到恶意投入问题不等,因此,使用DNS的网络物理系统可能会受到安全问题的困扰;近年来,出现了一个旨在解决这些安全问题的先进技术园区;这项工作提供了对这些安全问题的结构化和广泛的概览;我们首先确定了描述旨在探测、量化或减轻这些缺陷的研究活动的不足之处类别;我们的文件既涉及机器学习专家,也涉及安全工程师:前者可能获益于涵盖的广泛机器学习专题和关于近期方法局限性的讨论;后者可能深入了解现代ML方法的具体特点;我们还希望我们的贡献能促进关于ML系统及其如何相应推进现有方法的战略的讨论。