Organisms in nature have evolved to exhibit flexibility in face of changes to the environment and/or to themselves. Artificial neural networks (ANNs) have proven useful for controlling of artificial agents acting in environments. However, most ANN models used for reinforcement learning-type tasks have a rigid structure that does not allow for varying input sizes. Further, they fail catastrophically if inputs are presented in an ordering unseen during optimization. We find that these two ANN inflexibilities can be mitigated and their solutions are simple and highly related. For permutation invariance, no optimized parameters can be tied to a specific index of the input elements. For size invariance, inputs must be projected onto a common space that does not grow with the number of projections. Based on these restrictions, we construct a conceptually simple model that exhibit flexibility most ANNs lack. We demonstrate the model's properties on multiple control problems, and show that it can cope with even very rapid permutations of input indices, as well as changes in input size. Ablation studies show that is possible to achieve these properties with simple feedforward structures, but that it is much easier to optimize recurrent structures.
翻译:自然的有机体已经演变为在面对环境和/或自身变化时表现出灵活性。人工神经网络(ANNs)已证明对控制环境中的人工代理物有用。然而,用于强化学习型任务的大多数ANN模型结构僵硬,不允许不同的输入大小。此外,如果在优化过程中输入在无序的状态下出现,它们就会灾难性地失败。我们发现,这两个ANN的灵活度是可以减轻的,其解决方案是简单和高度相关的。对于变异,任何优化参数都无法与输入元素的具体指数挂钩。对于变异性,投入必须投射到一个不会随着预测数量增长而增长的共同空间。根据这些限制,我们建立一个概念上简单的模型,展示ANNs最缺乏的灵活性。我们展示了该模型在多个控制问题上的特性,并表明它甚至能够非常迅速地应对输入指数的变异异性,以及输入大小的变化。调整研究表明,用简单的进料结构实现这些属性是可能的,但最易优化的。