Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy that best fits observed sequences of states and actions implemented by an expert. Many algorithms for IRL have an inherently nested structure: the inner loop finds the optimal policy given parametrized rewards while the outer loop updates the estimates towards optimizing a measure of fit. For high dimensional environments such nested-loop structure entails a significant computational burden. To reduce the computational burden of a nested loop, novel methods such as SQIL [1] and IQ-Learn [2] emphasize policy estimation at the expense of reward estimation accuracy. However, without accurate estimated rewards, it is not possible to do counterfactual analysis such as predicting the optimal policy under different environment dynamics and/or learning new tasks. In this paper we develop a novel single-loop algorithm for IRL that does not compromise reward estimation accuracy. In the proposed algorithm, each policy improvement step is followed by a stochastic gradient step for likelihood maximization. We show that the proposed algorithm provably converges to a stationary solution with a finite-time guarantee. If the reward is parameterized linearly, we show the identified solution corresponds to the solution of the maximum entropy IRL problem. Finally, by using robotics control problems in MuJoCo and their transfer settings, we show that the proposed algorithm achieves superior performance compared with other IRL and imitation learning benchmarks.
翻译:反强化学习(IRL) 旨在恢复奖励功能以及最适合专家所实施的国家和行动所观察到的一系列国家和行动的配套最佳政策。 IRL的许多算法都有固有的嵌套结构: 内环会发现最佳政策, 给匹配的奖励, 而外环会更新估计数, 以优化适配度。 对于高维环境, 诸如嵌套环结构, 包含巨大的计算负担。 为了减少嵌套环的计算负担, SQIL [1] 和 IQ-Learn 等新方法强调政策估计, 以牺牲估计的准确性为代价。 但是, 没有准确的估计奖励, 就无法进行反事实分析, 如预测不同环境动态下的最佳政策和/或学习新任务。 在本文中, 我们为IRL制定新的单环算算法, 并不影响奖励估计的准确性。 在拟议的算法中, 每一个政策改进步骤都有一个随机梯梯梯梯梯步, 以便有可能实现最大化。 我们表明, 拟议的算法与固定的解决方案相匹配, 与固定时间比额的递定基准, 如果评级的系统化的系统化解决方案与最终显示, 我们的升级的升级的升级的升级的学习解决方案, 显示, 的升级的升级的升级的解决方案, 最后显示, 的升级的升级的升级的升级的解决方案与升级的升级的升级的解决方案与升级的升级的解决方案。