Keeping risk under control is often more crucial than maximizing expected rewards in real-world decision-making situations, such as finance, robotics, autonomous driving, etc. The most natural choice of risk measures is variance, which penalizes the upside volatility as much as the downside part. Instead, the (downside) semivariance, which captures the negative deviation of a random variable under its mean, is more suitable for risk-averse proposes. This paper aims at optimizing the mean-semivariance (MSV) criterion in reinforcement learning w.r.t. steady reward distribution. Since semivariance is time-inconsistent and does not satisfy the standard Bellman equation, the traditional dynamic programming methods are inapplicable to MSV problems directly. To tackle this challenge, we resort to Perturbation Analysis (PA) theory and establish the performance difference formula for MSV. We reveal that the MSV problem can be solved by iteratively solving a sequence of RL problems with a policy-dependent reward function. Further, we propose two on-policy algorithms based on the policy gradient theory and the trust region method. Finally, we conduct diverse experiments from simple bandit problems to continuous control tasks in MuJoCo, which demonstrate the effectiveness of our proposed methods.
翻译:控制风险往往比在金融、机器人、自主驾驶等现实决策情况下尽量扩大预期回报更为关键。 风险措施的最自然选择是差异,对上波动和下坡部分同样不利。相反,(下坡)半变量反映随机变量在其平均值下的负偏差,更适合风险反向提议。本文件旨在优化在强化学习中的平均偏差(MSV)标准,稳定奖赏分配。由于半偏差与时间不一致,不符合标准贝尔曼方程式,传统的动态方案拟订方法无法直接适用于MSV问题。为了应对这一挑战,我们采用Perturbation 分析(PA)理论,为MSV确定性能差异公式。我们指出,MSV问题可以通过以基于政策的奖赏功能迭代解决RL问题序列来解决。此外,我们根据政策梯度理论和信任区域方法提出了两种政策上的算法。最后,我们从简单的MuJoo系统问题到连续控制方法,我们进行了不同的方法的实验。</s>