Autonomous micromobility has been attracting the attention of researchers and practitioners in recent years. A key component of many micro-transport vehicles is the DC motor, a complex dynamical system that is continuous and non-linear. Learning to quickly control the DC motor in the presence of disturbances and uncertainties is desired for various applications that require robustness and stability. Techniques to accomplish this task usually rely on a mathematical system model, which is often insufficient to anticipate the effects of time-varying and interrelated sources of non-linearities. While some model-free approaches have been successful at the task, they rely on massive interactions with the system and are trained in specialized hardware in order to fit a highly parameterized controller. In this work, we learn to steer a DC motor via sample-efficient reinforcement learning. Using data collected from hardware interactions in the real world, we additionally build a simulator to experiment with a wide range of parameters and learning strategies. With the best parameters found, we learn an effective control policy in one minute and 53 seconds on a simulation and in 10 minutes and 35 seconds on a physical system.
翻译:近年来,自主微机动性一直引起研究人员和从业人员的注意,许多微型运输工具的一个关键组成部分是DC马达,这是一个连续和非线性的综合动态系统。在动荡和不确定的情况下学习迅速控制DC马达,对于各种需要稳健和稳定的应用程序来说,需要学习迅速控制DC马达。完成这项任务的技术通常依赖数学系统模型,该模型往往不足以预测非线性时间变化和相互关联来源的影响。虽然一些不使用模型的方法在任务中取得了成功,但它们依赖与系统的大规模互动,并受过专门硬件的培训,以适应高度参数化的控制器。在这项工作中,我们学习如何通过抽样高效的强化学习来引导DC马达。我们利用在现实世界中从硬件互动中收集的数据,另外建立一个模拟器,以试验范围广泛的参数和学习战略。有了最佳参数,我们在模拟过程中在1分53秒和10分35秒的物理系统中学习了有效的控制政策。