Feed-forward convolutional neural networks (CNNs) are currently state-of-the-art for object classification tasks such as ImageNet. Further, they are quantitatively accurate models of temporally-averaged responses of neurons in the primate brain's visual system. However, biological visual systems have two ubiquitous architectural features not shared with typical CNNs: local recurrence within cortical areas, and long-range feedback from downstream areas to upstream areas. Here we explored the role of recurrence in improving classification performance. We found that standard forms of recurrence (vanilla RNNs and LSTMs) do not perform well within deep CNNs on the ImageNet task. In contrast, novel cells that incorporated two structural features, bypassing and gating, were able to boost task accuracy substantially. We extended these design principles in an automated search over thousands of model architectures, which identified novel local recurrent cells and long-range feedback connections useful for object recognition. Moreover, these task-optimized ConvRNNs matched the dynamics of neural activity in the primate visual system better than feedforward networks, suggesting a role for the brain's recurrent connections in performing difficult visual behaviors.
翻译:Feed-forward Convolutional 神经网络(CNNs)目前是图像网络等物体分类任务的最新工艺。 此外,它们是原始大脑视觉系统中神经元反应时间平均反应的定量准确模型。然而,生物视觉系统有两个与典型CNN没有分享的无处不在的建筑特征:在皮层地区发生局部复发,以及下游地区的长距离反馈到上游地区。我们在这里探讨了重现在提高分类性能方面的作用。我们发现,在图像网络任务中,标准重现形式(Vanilla RNS和LSTMs)在深处CNNs中表现不佳。相比之下,包含两个结构特征(绕行和引线)的新细胞能够极大地提高任务准确性。我们在自动搜索数千个模型结构中扩展了这些设计原则,其中确定了新的本地经常细胞和长距离反馈连接对物体识别有用。此外,这些任务优化的CONRNNS与原始视觉系统中的神经活动动态相匹配,比向前向网络更好,表明在难以进行视觉联系的大脑经常性关系中的作用。