This paper presents an efficient procedure for multi-objective model checking of long-run average reward (aka: mean pay-off) and total reward objectives as well as their combination. We consider this for Markov automata, a compositional model that captures both traditional Markov decision processes (MDPs) as well as a continuous-time variant thereof. The crux of our procedure is a generalization of Forejt et al.'s approach for total rewards on MDPs to arbitrary combinations of long-run and total reward objectives on Markov automata. Experiments with a prototypical implementation on top of the Storm model checker show encouraging results for both model types and indicate a substantial improved performance over existing multi-objective long-run MDP model checking based on linear programming.
翻译:本文介绍了对长期平均报酬(aka: 平均报酬)和总奖赏目标及其组合进行多目标模式检查的高效程序。我们认为,对于Markov Automata来说,这是一个包含传统马尔科夫决策过程(MDPs)及其连续时间变式的构成模型。我们程序的核心是将Forejt等人(Forejt et al.)对MDP的全部奖赏方法加以概括,将其任意结合到Markov Automata的长期和全部奖赏目标。在暴风雨模式检查器上进行原型执行的实验显示两种模式都取得了令人鼓舞的结果,并表明与基于线性编程的现有多目标长期MDP模式检查相比,业绩有了显著改善。