We inspect the long-term learning ability of Long Short-Term Memory language models (LSTM LMs) by evaluating a contextual extension based on the Continuous Bag-of-Words (CBOW) model for both sentence- and discourse-level LSTM LMs and by analyzing its performance. We evaluate on text and speech. Sentence-level models using the long-term contextual module perform comparably to vanilla discourse-level LSTM LMs. On the other hand, the extension does not provide gains for discourse-level models. These findings indicate that discourse-level LSTM LMs already rely on contextual information to perform long-term learning.
翻译:我们检查长期短期记忆语言模型(LSTM LMs)的长期学习能力,方法是根据持续文字包模型(CBOW)的判刑和谈话级LSTM LMs(CBOW)模式评价一个背景扩展,并分析其表现;我们评估文本和演讲;使用长期背景模块的判刑级模型与香草级LSTM LMs的学习水平相当。另一方面,扩展并不为对话级模型带来收益。 这些调查结果表明,对话级LSTM LMs已经依靠背景信息进行长期学习。