We describe a new class of learning models called memory networks. Memory networks reason with inference components combined with a long-term memory component; they learn how to use these jointly. The long-term memory can be read and written to, with the goal of using it for prediction. We investigate these models in the context of question answering (QA) where the long-term memory effectively acts as a (dynamic) knowledge base, and the output is a textual response. We evaluate them on a large-scale QA task, and a smaller, but more complex, toy task generated from a simulated world. In the latter, we show the reasoning power of such models by chaining multiple supporting sentences to answer questions that require understanding the intension of verbs.
翻译:我们描述一个新的学习模式类别,称为记忆网络。记忆网络与推论组成部分结合一个长期记忆组成部分;它们学会如何共同使用这些组成部分;长期记忆可以读写,目的是用它来预测。我们在回答问题(QA)时调查这些模型,长期记忆有效地作为(动态)知识基础,而产出是文字反应。我们根据一个大规模QA任务来评估这些模型,以及一个由模拟世界产生的较小但更复杂的玩具任务。在后一种世界中,我们通过将多个支持性句子连锁起来回答需要理解动词强度的问题,来显示这些模型的推理力。