Central to the design of many robot systems and their controllers is solving a constrained blackbox optimization problem. This paper presents CNMA, a new method of solving this problem that is conservative in the number of potentially expensive blackbox function evaluations; allows specifying complex, even recursive constraints directly rather than as hard-to-design penalty or barrier functions; and is resilient to the non-termination of function evaluations. CNMA leverages the ability of neural networks to approximate any continuous function, their transformation into equivalent mixed integer linear programs (MILPs) and their optimization subject to constraints with industrial strength MILP solvers. A new learning-from-failure step guides the learning to be relevant to solving the constrained optimization problem. Thus, the amount of learning is orders of magnitude smaller than that needed to learn functions over their entire domains. CNMA is illustrated with the design of several robotic systems: wave-energy propelled boat, lunar lander, hexapod, cartpole, acrobot and parallel parking. These range from 6 real-valued dimensions to 36. We show that CNMA surpasses the Nelder-Mead, Gaussian and Random Search optimization methods against the metric of number of function evaluations.
翻译:许多机器人系统及其控制器的设计中心正在解决一个限制的黑盒优化问题。 本文展示了CNMA, 这是一种解决这一问题的新方法,在潜在昂贵黑盒功能评估的数量上是保守的; 允许直接具体说明复杂、甚至循环的制约,而不是难以设计的惩罚或屏障功能; 并且能够适应功能评估的不终结。 CNMA利用神经网络的能力来接近任何连续功能,将其转化成等同的混合整形线性程序(MILP),并优化,但受工业实力MILP解决方案的限制。 一个新的从失败中学习的步骤引导学习与解决限制的优化问题相关。 因此,学习的数量比学习其整个领域功能所需的数量小。 CNMA通过设计若干机器人系统来加以说明:波能驱动船、月球登陆器、六极、马波德、马波尔波尔特、一个crobot和平行停车处。 这些系统从6个实际价值层面到36个层面。 我们显示CNMA超越了Nlder-Mead, Gaus 和随机优化方法的数量。