As large pre-trained image-processing neural networks are being embedded in autonomous agents such as self-driving cars or robots, the question arises of how such systems can communicate with each other about the surrounding world, despite their different architectures and training regimes. As a first step in this direction, we systematically explore the task of referential communication in a community of state-of-the-art pre-trained visual networks, showing that they can develop a shared protocol to refer to a target image among a set of candidates. Such shared protocol, induced in a self-supervised way, can to some extent be used to communicate about previously unseen object categories, as well as to make more granular distinctions compared to the categories taught to the original networks. Contradicting a common view in multi-agent emergent communication research, we find that imposing a discrete bottleneck on communication hampers the emergence of a general code. Moreover, we show that a new neural network can learn the shared protocol developed in a community with remarkable ease, and the process of integrating a new agent into a community more stably succeeds when the original community includes a larger set of heterogeneous networks. Finally, we illustrate the independent benefits of developing a shared communication layer by using it to directly transfer an object classifier from a network to another, and we qualitatively and quantitatively study its emergent properties.
翻译:由于经过预先训练的大型图像处理神经网络正在嵌入自驾驶汽车或机器人等自主代理器中,因此产生了这样一个问题:尽管这些系统的结构和培训制度各不相同,它们如何能够就周围的世界相互交流。作为朝这个方向迈出的第一步,我们系统地探索在一个最先进的先受过训练的视觉网络社区中进行特惠通信的任务,表明它们可以开发一个共同的协议,在一组候选人中提及一个目标图像。这种以自我监督的方式引导的共享协议,可以在某种程度上用于交流以前看不见的物体类别,并比最初的网络所教授的类别作出更多的粒子区分。在多剂突发通信研究中,我们发现对通信设置一个离散的瓶颈会妨碍形成一个通用代码。此外,我们表明一个新的神经网络可以很容易地学习在社区中开发的共享协议,并且将新的代理器融入社区的过程,在最初的社区包括一系列较大型的混合网络之后,可以更加精确地成功地进行联系。最后,我们通过一个从一个更大规模的混合的网络到一个不断升级的层次的研究,我们通过另一个层次来展示了另一个独立的传播的层次的特性。