For a receding-horizon controller with a known system and with an approximate terminal value function, it is well-known that increasing the prediction horizon can improve its control performance. However, when the prediction model is inexact, a larger prediction horizon also causes propagation and accumulation of the prediction error. In this work, we aim to analyze the effect of the above trade-off between the modeling error, the terminal value function error, and the prediction horizon on the performance of a nominal receding-horizon linear quadratic (LQ) controller. By developing a novel perturbation result of the Riccati difference equation, a performance upper bound is obtained and suggests that for many cases, the prediction horizon should be either 1 or infinity to improve the control performance, depending on the relative difference between the modeling error and the terminal value function error. The obtained suboptimality performance bound is also applied to provide end-to-end performance guarantees, e.g., regret bounds, for nominal receding-horizon LQ controllers in a learning-based setting.
翻译:对于具有已知系统和近似终端值功能的递减正数控制器,众所周知,增加预测地平线可以改善其控制性能。然而,当预测模型不精确时,更大的预测地平线也会导致预测误差的传播和累积。在这项工作中,我们的目标是分析上述模型错误、终端值函数错误和名义递减偏差线性二次曲线控制器性能的预测地平线。通过开发Riccati差异方程式的新的扰动结果,获得一个性能上限,并表明在许多情况下,根据模型错误和终端值函数错误之间的相对差异,预测地平线应改进控制性能。获得的次优性能约束还用于在学习环境中为名义递减-Horizon LQ控制器提供端到端性性性能保证,例如,遗憾约束。