Multi-objective reinforcement learning (MORL) approaches have emerged to tackle many real-world problems with multiple conflicting objectives by maximizing a joint objective function weighted by a preference vector. These approaches find fixed customized policies corresponding to preference vectors specified during training. However, the design constraints and objectives typically change dynamically in real-life scenarios. Furthermore, storing a policy for each potential preference is not scalable. Hence, obtaining a set of Pareto front solutions for the entire preference space in a given domain with a single training is critical. To this end, we propose a novel MORL algorithm that trains a single universal network to cover the entire preference space scalable to continuous robotic tasks. The proposed approach, Preference-Driven MORL (PD-MORL), utilizes the preferences as guidance to update the network parameters. It also employs a novel parallelization approach to increase sample efficiency. We show that PD-MORL achieves up to 25% larger hypervolume for challenging continuous control tasks and uses an order of magnitude fewer trainable parameters compared to prior approaches.
翻译:多个目标强化学习(MORL)方法已经出现,通过最大限度地提高一个由偏好矢量加权的共同目标功能,解决许多现实世界问题,实现多个相互冲突的目标。这些方法发现与培训期间指定的偏爱矢量相对应的固定定制政策。然而,设计限制和目标通常在现实情景中动态变化。此外,为每个潜在偏爱存储一项政策是不可扩展的。因此,为特定域的全部偏爱空间获得一套有单一培训的Pareto前方解决方案至关重要。为此,我们提出一个新的MORL算法,用于培训一个单一的通用网络,以覆盖整个可扩展至连续机器人任务的偏爱空间。拟议方法Preg-Driven MORL(PD-MORL)将偏好选项用作更新网络参数的指南。它还采用新的平行方法提高样本效率。我们显示,PD-MORL在挑战持续控制任务方面达到高达25%的超容量,并使用比以往方法少得多的可培训参数。