Learning controller parameters from closed-loop data has been shown to improve closed-loop performance. Bayesian optimization, a widely used black-box and sample-efficient learning method, constructs a probabilistic surrogate of the closed-loop performance from few experiments and uses it to select informative controller parameters. However, it typically struggles with dense high-dimensional controller parameterizations, as they may appear, for example, in tuning model predictive controllers, because standard surrogate models fail to capture the structure of such spaces. This work suggests that the use of Bayesian neural networks as surrogate models may help to mitigate this limitation. Through a comparison between Gaussian processes with Matern kernels, finite-width Bayesian neural networks, and infinite-width Bayesian neural networks on a cart-pole task, we find that Bayesian neural network surrogate models achieve faster and more reliable convergence of the closed-loop cost and enable successful optimization of parameterizations with hundreds of dimensions. Infinite-width Bayesian neural networks also maintain performance in settings with more than one thousand parameters, whereas Matern-kernel Gaussian processes rapidly lose effectiveness. These results indicate that Bayesian neural network surrogate models may be suitable for learning dense high-dimensional controller parameterizations and offer practical guidance for selecting surrogate models in learning-based controller design.
翻译:从闭环数据中学习控制器参数已被证明能够提升闭环性能。贝叶斯优化作为一种广泛使用的黑盒且样本高效的学习方法,通过少量实验构建闭环性能的概率代理模型,并利用其选择信息丰富的控制器参数。然而,该方法通常难以处理密集的高维控制器参数化问题,例如在调整模型预测控制器时可能出现的情况,因为标准代理模型难以捕捉此类空间的结构。本研究提出,采用贝叶斯神经网络作为代理模型可能有助于缓解这一局限。通过在倒立摆任务中比较采用Matern核的高斯过程、有限宽度贝叶斯神经网络和无限宽度贝叶斯神经网络,我们发现贝叶斯神经网络代理模型能够实现闭环成本更快、更可靠的收敛,并成功优化数百维的参数化配置。无限宽度贝叶斯神经网络在参数超过一千个的场景下仍能保持性能,而采用Matern核的高斯过程则迅速失效。这些结果表明,贝叶斯神经网络代理模型可能适用于学习密集的高维控制器参数化,并为基于学习的控制器设计中代理模型的选择提供了实用指导。