Autonomous systems with machine learning-based perception can exhibit unpredictable behaviors that are difficult to quantify, let alone verify. Such behaviors are convenient to capture in probabilistic models, but probabilistic model checking of such models is difficult to scale -- largely due to the non-determinism added to models as a prerequisite for provable conservatism. Statistical model checking (SMC) has been proposed to address the scalability issue. However it requires large amounts of data to account for the aforementioned non-determinism, which in turn limits its scalability. This work introduces a general technique for reduction of non-determinism based on assumptions of "monotonic safety'", which define a partial order between system states in terms of their probabilities of being safe. We exploit these assumptions to remove non-determinism from controller/plant models to drastically speed up probabilistic model checking and statistical model checking while providing provably conservative estimates as long as the safety is indeed monotonic. Our experiments demonstrate model-checking speed-ups of an order of magnitude while maintaining acceptable accuracy and require much less data for accurate estimates when running SMC -- even when monotonic safety does not perfectly hold and provable conservatism is not achieved.
翻译:以机器学习为基础的自动系统可能会出现难以量化的不可预测行为,更不用说核实了。这种行为很容易在概率模型中捕捉到,但这种模型的概率模型检查很难规模化 -- -- 主要是因为模型中增加了非确定性,作为可变保守主义的先决条件。提议进行统计模型检查(SMC)是为了解决可缩放性问题。然而,它需要大量的数据来说明上述非确定性,而这反过来又限制了其可缩放性。这项工作引入了一种基于“分子安全”假设的减少非确定性的一般技术,该技术界定了系统各州之间在安全概率方面的部分顺序。我们利用这些假设来消除控制/工厂模型中的非确定性,以大大加快概率模型检查和统计模型检查的速度,同时提供可靠保守性的估计数,因为安全性的确是单调的。我们的实验表明,在进行精确性估算时,对数量级的加速度进行模型检查,同时保持可接受性准确性的准确性,要求系统国家之间的不那么精确性数据。