This note addresses the problem of evaluating the impact of an attack on discrete-time nonlinear stochastic control systems. The problem is formulated as an optimal control problem with a joint chance constraint that forces the adversary to avoid detection throughout a given time period. Due to the joint constraint, the optimal control policy depends not only on the current state, but also on the entire history, leading to an explosion of the search space and making the problem generally intractable. However, we discover that the current state and whether an alarm has been triggered, or not, is sufficient for specifying the optimal decision at each time step. This information, which we refer to as the alarm flag, can be added to the state space to create an equivalent optimal control problem that can be solved with existing numerical approaches using a Markov policy. Additionally, we note that the formulation results in a policy that does not avoid detection once an alarm has been triggered. We extend the formulation to handle multi-alarm avoidance policies for more reasonable attack impact evaluations, and show that the idea of augmenting the state space with an alarm flag is valid in this extended formulation as well.
翻译:本说明涉及评估攻击离散时间非线性随机控制系统的影响的问题。 这个问题被作为一种最佳控制问题, 并带有共同的机会限制, 迫使对手在整个特定时期内避免被探测。 由于共同的限制, 最佳控制政策不仅取决于当前状况, 也取决于整个历史, 导致搜索空间爆炸, 使问题普遍变得棘手。 然而, 我们发现, 目前的状态, 以及是否已经触发警报, 足以在每一个步骤中具体确定最佳决定。 我们称之为警报旗的信息, 可以添加到州空间, 以形成一个相当的最佳控制问题, 通过使用马尔科夫政策的现有数字方法解决。 此外, 我们注意到, 制定政策的结果并不避免在警报发生后被探测。 我们扩展了设计, 处理多报警避免政策, 以便进行更合理的攻击影响评估, 并表明, 以警报旗增强国家空间的想法在这种扩展的表述中也是有效的。