Actors and critics in actor-critic reinforcement learning algorithms are functionally separate, yet they often use the same network architectures. This case study explores the performance impact of network sizes when considering actor and critic architectures independently. By relaxing the assumption of architectural symmetry, it is often possible for smaller actors to achieve comparable policy performance to their symmetric counterparts. Our experiments show up to 97% reduction in the number of network weights with an average reduction of 64% over multiple algorithms on multiple tasks. Given the practical benefits of reducing actor complexity, we believe configurations of actors and critics are aspects of actor-critic design that deserve to be considered independently.
翻译:行为体强化学习算法的参与者和批评者在功能上是分开的,但他们经常使用相同的网络架构。 本案例研究在独立考虑行为体和评论者架构时,探讨了网络规模的绩效影响。 通过放松对建筑对称的假设,较小行为体往往有可能实现与其对称对应方相似的政策绩效。 我们的实验显示,网络加权数减少高达97%,比多重任务多重算法平均减少64%。 考虑到降低行为体复杂性的实际好处,我们认为行为体和批评者构成的构成是行为体-批评设计中值得独立考虑的方面。