We present Distributed Deep Deterministic Policy Gradient (3DPG), a multi-agent actor-critic (MAAC) algorithm for Markov games. Unlike previous MAAC algorithms, 3DPG is fully distributed during both training and deployment. 3DPG agents calculate local policy gradients based on the most recently available local data (states, actions) and local policies of other agents. During training, this information is exchanged using a potentially lossy and delaying communication network. The network therefore induces Age of Information (AoI) for data and policies. We prove the asymptotic convergence of 3DPG even in the presence of potentially unbounded Age of Information (AoI). This provides an important step towards practical online and distributed multi-agent learning since 3DPG does not assume information to be available deterministically. We analyze 3DPG in the presence of policy and data transfer under mild practical assumptions. Our analysis shows that 3DPG agents converge to a local Nash equilibrium of Markov games in terms of utility functions expressed as the expected value of the agents local approximate action-value functions (Q-functions). The expectations of the local Q-functions are with respect to limiting distributions over the global state-action space shaped by the agents' accumulated local experiences. Our results also shed light on the policies obtained by general MAAC algorithms. We show through a heuristic argument and numerical experiments that 3DPG improves convergence over previous MAAC algorithms that use old actions instead of old policies during training. Further, we show that 3DPG is robust to AoI; it learns competitive policies even with large AoI and low data availability.
翻译:3DPG 代理商根据最近可获得的地方数据(国家、行动)和其他地方代理商的地方政策计算地方政策梯度。在培训过程中,这种信息交流使用潜在损失和延迟的通信网络进行。因此,网络为数据和政策带来信息时代(AoI),我们证明3DPG 即使在可能不受约束的信息时代(AoI)存在的情况下,也无济于事。这提供了在培训和部署期间充分分布的3DPG 算法。3DPG 代理商根据最近可获得的地方数据(国家、行动)和其他地方代理商的地方政策计算地方政策梯度。我们的分析表明,3DPG 代理商通过可能丢失和延误的通信网络,在数据和政策方面,为信息时代(AoI)带来了信息时代(Ao)的不协调。我们证明,3DPG 即使在可能不受约束的信息时代约束的信息时代(AoI ) 也为实用的在线和分布提供了重要的多机构学习步骤。我们通过以往的A-DA-DA 分析工具来改变老的老的老的动作。