During the last years, there has been a lot of interest in achieving some kind of complex reasoning using deep neural networks. To do that, models like Memory Networks (MemNNs) have combined external memory storages and attention mechanisms. These architectures, however, lack of more complex reasoning mechanisms that could allow, for instance, relational reasoning. Relation Networks (RNs), on the other hand, have shown outstanding results in relational reasoning tasks. Unfortunately, their computational cost grows quadratically with the number of memories, something prohibitive for larger problems. To solve these issues, we introduce the Working Memory Network, a MemNN architecture with a novel working memory storage and reasoning module. Our model retains the relational reasoning abilities of the RN while reducing its computational complexity from quadratic to linear. We tested our model on the text QA dataset bAbI and the visual QA dataset NLVR. In the jointly trained bAbI-10k, we set a new state-of-the-art, achieving a mean error of less than 0.5%. Moreover, a simple ensemble of two of our models solves all 20 tasks in the joint version of the benchmark.
翻译:在过去几年里,人们对于利用深层神经网络实现某种复杂的推理非常感兴趣,为此,记忆网络(MemNNs)等模型结合了外部记忆存储和关注机制。然而,这些结构缺乏更复杂的推理机制,因此,例如,关系推理机制。另一方面,关系网络(RNs)在关系推理任务方面显示出突出的结果。不幸的是,它们的计算成本随着记忆数量而四倍增长,对于更大的问题来说是令人望而却步。为了解决这些问题,我们引入了工作记忆网络,这是MemNN(MemN)结构,有一个创新的工作记忆存储和推理模块。我们的模型保留了RN的关系推理能力,同时将其计算复杂性从四边形降低到线形。我们用文本QA数据集bAbI和视觉QA数据集NLVR测试了我们的模型模型。在经过联合培训的bI-10k中,我们设置了一个新的状态,即实现低于0.5 %的平均错误。此外,我们的两个模型共同解决了两个版本的基准。