As machine learning (ML) being applied to many mission-critical scenarios, certifying ML model robustness becomes increasingly important. Many previous works focuses on the robustness of independent ML and ensemble models, and can only certify a very small magnitude of the adversarial perturbation. In this paper, we take a different viewpoint and improve learning robustness by going beyond independent ML and ensemble models. We aim at promoting the generic Sensing-Reasoning machine learning pipeline which contains both the sensing (e.g. deep neural networks) and reasoning (e.g. Markov logic networks (MLN)) components enriched with domain knowledge. Can domain knowledge help improve learning robustness? Can we formally certify the end-to-end robustness of such an ML pipeline? We first theoretically analyze the computational complexity of checking the provable robustness in the reasoning component. We then derive the provable robustness bound for several concrete reasoning components. We show that for reasoning components such as MLN and a specific family of Bayesian networks it is possible to certify the robustness of the whole pipeline even with a large magnitude of perturbation which cannot be certified by existing work. Finally, we conduct extensive real-world experiments on large scale datasets to evaluate the certified robustness for Sensing-Reasoning ML pipelines.
翻译:随着机器学习(ML)应用于许多任务关键情景,认证ML模型的稳健性变得日益重要。许多以前的工作都侧重于独立的ML和组合模型的稳健性,只能证明极小的对抗性扰动。在本文件中,我们采取不同的观点,通过超越独立的ML和组合模型来提高学习的稳健性。我们的目标是促进通用遥感-再协调机学习管道,该管道既包含遥感(例如深神经网络)组成部分,也包含推理(例如Markov逻辑网络(MLN)),以域知识丰富。域域知识能帮助提高学习稳健性吗?我们能否正式证明这种ML管道的末端到端的稳健性?我们首先从理论上分析检查推理部分中可确认的稳健性的计算复杂性。然后我们为若干具体推理组成部分找到可证实的稳性。我们表明,对于诸如MLN和Bayesian网络的特定组别(例如Markov逻辑网络)的构件,它能够证明整个管道的稳健性,甚至可以用大规模的轨迹来证明整个输油管的稳性,而我们最终无法通过验证我们已验证的大规模的大规模的大规模的实验室实验进行。