The ability to recover from an unexpected external perturbation is a fundamental motor skill in bipedal locomotion. An effective response includes the ability to not just recover balance and maintain stability but also to fall in a safe manner when balance recovery is physically infeasible. For robots associated with bipedal locomotion, such as humanoid robots and assistive robotic devices that aid humans in walking, designing controllers which can provide this stability and safety can prevent damage to robots or prevent injury related medical costs. This is a challenging task because it involves generating highly dynamic motion for a high-dimensional, non-linear and under-actuated system with contacts. Despite prior advancements in using model-based and optimization methods, challenges such as requirement of extensive domain knowledge, relatively large computational time and limited robustness to changes in dynamics still make this an open problem. In this thesis, to address these issues we develop learning-based algorithms capable of synthesizing push recovery control policies for two different kinds of robots : Humanoid robots and assistive robotic devices that assist in bipedal locomotion. Our work can be branched into two closely related directions : 1) Learning safe falling and fall prevention strategies for humanoid robots and 2) Learning fall prevention strategies for humans using a robotic assistive devices. To achieve this, we introduce a set of Deep Reinforcement Learning (DRL) algorithms to learn control policies that improve safety while using these robots.
翻译:从出乎意料的外部扰动中恢复的能力,是双叶动动动中的一项基本运动技能。有效的反应包括不仅能够恢复平衡并维持稳定,而且能够在实际无法实现平衡恢复时以安全的方式下降。对于与双叶动动动能相关的机器人,如人造机器人和辅助机器人装置,帮助人类步行,设计能够提供这种稳定性和安全的控制器,可以防止机器人受损或防止与伤害有关的医疗成本。这是一项具有挑战性的任务,因为它涉及为高维、非线性和低活性接触的系统产生高度动态的动态运动。尽管在使用基于模型和优化的方法方面已经取得了进步,但是在使用广域知识的需求、相对庞大的计算时间和对动态变化的强力有限等挑战仍然使这一问题成为公开的问题。为了解决这些问题,我们开发基于学习的算法,能够对两种不同的机器人的推力控制政策进行合成:人类机器人机器人机器人机器人机器人机器人机器人机器人和辅助机械机械机械装置,有助于双向移动。我们的工作可以进行双向地学习,同时学习机器人学习机器人的学习策略。