Multiagent systems provide an ideal environment for the evaluation and analysis of real-world problems using reinforcement learning algorithms. Most traditional approaches to multiagent learning are affected by long training periods as well as high computational complexity. NEAT (NeuroEvolution of Augmenting Topologies) is a popular evolutionary strategy used to obtain the best performing neural network architecture often used to tackle optimization problems in the field of artificial intelligence. This paper utilizes the NEAT algorithm to achieve competitive multiagent learning on a modified pong game environment in an efficient manner. The competing agents abide by different rules while having similar observation space parameters. The proposed algorithm utilizes this property of the environment to define a singular neuroevolutionary procedure that obtains the optimal policy for all the agents. The compiled results indicate that the proposed implementation achieves ideal behaviour in a very short training period when compared to existing multiagent reinforcement learning models.
翻译:多试剂系统为利用强化学习算法评价和分析现实世界问题提供了一个理想的环境。多试剂学习的大多数传统方法都受到长期培训期和高计算复杂性的影响。NEAT(增强地形学的神经进化)是一种流行的进化战略,用于获取最有效果的神经网络结构,通常用于解决人工智能领域的优化问题。本文件利用NEAT算法,在改制的海绵游戏环境中以有效的方式实现竞争性多试剂学习。竞争者遵守不同的规则,同时拥有类似的观测空间参数。拟议的算法利用环境的这种特性来界定一种单一的神经进化程序,为所有代理者获得最佳政策。汇编的结果表明,与现有的多试剂强化学习模型相比,拟议的实施在非常短的培训期内实现了理想的行为。