Statistical model checking is a class of sequential algorithms that can verify specifications of interest on an ensemble of cyber-physical systems (e.g., whether 99% of cars from a batch meet a requirement on their energy efficiency). These algorithms infer the probability that given specifications are satisfied by the systems with provable statistical guarantees by drawing sufficient numbers of independent and identically distributed samples. During the process of statistical model checking, the values of the samples (e.g., a user's car energy efficiency) may be inferred by intruders, causing privacy concerns in consumer-level applications (e.g., automobiles and medical devices). This paper addresses the privacy of statistical model checking algorithms from the point of view of differential privacy. These algorithms are sequential, drawing samples until a condition on their values is met. We show that revealing the number of the samples drawn can violate privacy. We also show that the standard exponential mechanism that randomizes the output of an algorithm to achieve differential privacy fails to do so in the context of sequential algorithms. Instead, we relax the conservative requirement in differential privacy that the sensitivity of the output of the algorithm should be bounded to any perturbation for any data set. We propose a new notion of differential privacy which we call expected differential privacy. Then, we propose a novel expected sensitivity analysis for the sequential algorithm and proposed a corresponding exponential mechanism that randomizes the termination time to achieve the expected differential privacy. We apply the proposed mechanism to statistical model checking algorithms to preserve the privacy of the samples they draw. The utility of the proposed algorithm is demonstrated in a case study.
翻译:统计模型检查是一系列顺序算法,可以核实一系列网络-物理系统(例如,99%的汽车是否满足其能源效率要求)的兴趣规格。这些算法推断出,特定规格由具有可证实的统计保证的系统满足的可能性,方法是通过抽取足够数量的独立和同样分布的样本。在统计模型检查过程中,抽样的价值(例如用户的汽车能源效率)可由入侵者推断,从而引起消费者一级应用(例如汽车和医疗设备)的隐私问题。本文从差异隐私权的角度处理统计模型检查算法的隐私问题。这些算法按顺序排列,在满足其价值的条件之前绘制样本。我们表明,披露所抽取的样本数量会侵犯隐私。我们还表明,在序列算法中,随机抽调算算算算算算算算算算算得出不同隐私的输出结果时,在消费者一级应用的隐私(例如汽车和医疗设备)中引起隐私关切。本文从差异隐私的角度论述统计模型检查算算算法的隐私隐私权。这些算法是顺序顺序的隐私的隐私隐私隐私隐私隐私隐私隐私隐私隐私隐私隐私隐私隐私隐私,我们提出一个预期的敏感度,然后提出一个预期的序列分析,我们提出一个预期的顺序分析,我们提出一个新的分析,然后提出一个新的分析。我们提出一个预期的精确分析,然后提出一个新的分析。我们提出一个新的分析。我们提出一个新的分析。我们提出一个预期的精确分析。我们提出一个新的分析。我们提出一个新的分析。我们提出一个新的分析。