Feed-forward convolutional neural networks (CNNs) are currently state-of-the-art for object classification tasks such as ImageNet. Further, they are quantitatively accurate models of temporally-averaged responses of neurons in the primate brain's visual system. However, biological visual systems have two ubiquitous architectural features not shared with typical CNNs: local recurrence within cortical areas, and long-range feedback from downstream areas to upstream areas. Here we explored the role of recurrence in improving classification performance. We found that standard forms of recurrence (vanilla RNNs and LSTMs) do not perform well within deep CNNs on the ImageNet task. In contrast, custom cells that incorporated two structural features, bypassing and gating, were able to boost task accuracy substantially. We extended these design principles in an automated search over thousands of model architectures, which identified novel local recurrent cells and long-range feedback connections useful for object recognition. Moreover, these task-optimized ConvRNNs explained the dynamics of neural activity in the primate visual system better than feedforward networks, suggesting a role for the brain's recurrent connections in performing difficult visual behaviors.
翻译:Feed-forward Convolutional 神经网络(CNNs)目前是图像网络等物体分类任务的最新工艺。 此外,它们是原始大脑视觉系统中神经元反应时间平均反应的定量准确模型。然而,生物视觉系统有两个与典型CNN没有分享的无处不在的建筑特征:在皮层地区发生局部复发,以及下游地区的长距离反馈到上游地区。我们在这里探讨了重现在提高分类性能方面的作用。我们发现,在图像网络任务中,标准重现形式(Vanilla RNS和LSTMs)在深处CNNs中表现不佳。相比之下,包含两个结构特征的定制细胞绕过和粘贴能够极大地提高任务准确性。我们在自动搜索数千个模型结构时扩展了这些设计原则,这些模型确定了新的本地经常细胞和长距离反馈连接对物体识别有用。此外,这些任务优化的CONRNNS还解释了原始视觉系统中神经活动的动态比向前网络更好,表明在难以视觉关系中进行视觉反应的角色。