Existing reasoning tasks often have an important assumption that the input contents can be always accessed while reasoning, requiring unlimited storage resources and suffering from severe time delay on long sequences. To achieve efficient reasoning on long sequences with limited storage resources, memory augmented neural networks introduce a human-like write-read memory to compress and memorize the long input sequence in one pass, trying to answer subsequent queries only based on the memory. But they have two serious drawbacks: 1) they continually update the memory from current information and inevitably forget the early contents; 2) they do not distinguish what information is important and treat all contents equally. In this paper, we propose the Rehearsal Memory (RM) to enhance long-sequence memorization by self-supervised rehearsal with a history sampler. To alleviate the gradual forgetting of early information, we design self-supervised rehearsal training with recollection and familiarity tasks. Further, we design a history sampler to select informative fragments for rehearsal training, making the memory focus on the crucial information. We evaluate the performance of our rehearsal memory by the synthetic bAbI task and several downstream tasks, including text/video question answering and recommendation on long sequences.
翻译:现有的推理任务往往有一个重要假设,即输入内容在推理过程中总是可以得到,需要无限的储存资源,而且长时间的顺序有严重拖延。为了在储存资源有限的情况下对长序列进行有效的推理,内存增强神经网络引入了一种人式的读写记忆,以压缩和将长输入序列在一关内记忆,试图只根据记忆来回答随后的询问。但是,它们有两个严重的缺点:(1)它们不断从当前信息中更新记忆,不可避免地忘记早期内容;(2)它们不区分什么信息是重要的,并且对所有内容一视同仁。在本文件中,我们建议重新演练记忆(RM)通过对历史采样者进行自我监督的演练,加强长期序列的记忆化。为了减轻早期信息的逐渐遗忘,我们设计了以记忆和熟悉任务为基础自我监督的演练训练。此外,我们设计了一个历史采样器,为排练培训选择信息性碎片,将记忆的焦点放在关键信息上。我们通过合成的BABI任务和若干下游任务,包括文字/录像问题解答和关于长期序列的建议,评估我们的演练记忆的绩效表现。