Artificial neural systems trained using reinforcement, supervised, and unsupervised learning all acquire internal representations of high dimensional input. To what extent these representations depend on the different learning objectives is largely unknown. Here we compare the representations learned by eight different convolutional neural networks, each with identical ResNet architectures and trained on the same family of egocentric images, but embedded within different learning systems. Specifically, the representations are trained to guide action in a compound reinforcement learning task; to predict one or a combination of three task-related targets with supervision; or using one of three different unsupervised objectives. Using representational similarity analysis, we find that the network trained with reinforcement learning differs most from the other networks. Using metrics inspired by the neuroscience literature, we find that the model trained with reinforcement learning has a sparse and high-dimensional representation wherein individual images are represented with very different patterns of neural activity. Further analysis suggests these representations may arise in order to guide long-term behavior and goal-seeking in the RL agent. Finally, we compare the representations learned by the RL agent to neural activity from mouse visual cortex and find it to perform as well or better than other models. Our results provide insights into how the properties of neural representations are influenced by objective functions and can inform transfer learning approaches.
翻译:利用强化、监督和不受监督的学习,经过培训的人工神经系统都获得高维投入的内部表现。 这些表现在多大程度上取决于不同的学习目标,基本上不为人所知。 在这里,我们比较了八个不同的进化神经网络所学的表述,每个网络都具有相同的ResNet结构,并受过相同的自我中心图像家庭的培训,但都嵌入不同的学习系统。具体地说,这些表述经过培训,以指导复合强化学习任务的行动;预测一个或三个任务相关目标与监督的组合;或使用三个不同但不受监督的目标之一。使用代表性相似性分析,我们发现受过强化学习培训的网络与其他网络相比差别最大。我们利用神经科学文献所启发的衡量标准,我们发现受过强化学习培训的模型具有稀疏和高维度的表述,其中个人图像具有非常不同的神经活动模式。进一步分析表明,这些表述可能出现于指导长期行为和在RL代理中追求目标。最后,我们比较了RL代理所学到的演示,从鼠视内特质活动中获得的神经活动,我们通过客观的思维方式可以更好地了解我们的视觉表现。