The unaffordable computation load of nonlinear model predictive control (NMPC) has prevented it for being used in robots with high sampling rates for decades. This paper is concerned with the policy learning problem for nonlinear MPC with system constraints, and its applications to unmanned surface vehicles (USVs), where the nonlinear MPC policy is learned offline and deployed online to resolve the computational complexity issue. A deep neural networks (DNN) based policy learning MPC (PL-MPC) method is proposed to avoid solving nonlinear optimal control problems online. The detailed policy learning method is developed and the PL-MPC algorithm is designed. The strategy to ensure the practical feasibility of policy implementation is proposed, and it is theoretically proved that the closed-loop system under the proposed method is asymptotically stable in probability. In addition, we apply the PL-MPC algorithm successfully to the motion control of USVs. It is shown that the proposed algorithm can be implemented at a sampling rate up to $5 Hz$ with high-precision motion control. The experiment video is available via:\url{https://v.youku.com/v_show/id_XNTkwMTM0NzM5Ng==.html
翻译:非线性模型预测控制(NMPC)无法负担的计算载荷使得无法在数十年来用于具有高取样率的非线性模型预测控制(NMPC)的机器人中使用。本文涉及非线性MPC在系统限制下的政策学习问题及其对无人驾驶的表面飞行器(USVs)的应用问题,因为非线性MPC政策是离线性的,并在线部署以解决计算复杂性问题。提出了基于深度神经网络的政策学习 MPC (DNN) 方法,以避免在网上解决非线性最佳控制问题。详细的政策学习方法和PL-MPC 算法已经制定。提出了确保政策执行实际可行性的战略,并在理论上证明,拟议方法下的闭环系统很可能是无线性稳定的。此外,我们成功地将PL-MPC算法应用于USVs的运动控制。已经表明,拟议的算法可以以高达5Hz$的采样率在高精确度动作控制下实施。实验视频可通过以下途径获得:\url_NHMDF_M5_MKV.Mvvvvks.