Although neural language models are effective at capturing statistics of natural language, their representations are challenging to interpret. In particular, it is unclear how these models retain information over multiple timescales. In this work, we construct explicitly multi-timescale language models by manipulating the input and forget gate biases in a long short-term memory (LSTM) network. The distribution of timescales is selected to approximate power law statistics of natural language through a combination of exponentially decaying memory cells. We then empirically analyze the timescale of information routed through each part of the model using word ablation experiments and forget gate visualizations. These experiments show that the multi-timescale model successfully learns representations at the desired timescales, and that the distribution includes longer timescales than a standard LSTM. Further, information about high-,mid-, and low-frequency words is routed preferentially through units with the appropriate timescales. Thus we show how to construct language models with interpretable representations of different information timescales.
翻译:虽然神经语言模型在获取自然语言统计数据方面是有效的,但其表达方式是难以解释的。特别是,不清楚这些模型如何在多个时间尺度内保留信息。在这项工作中,我们通过在长期短期内存(LSTM)网络中操纵输入并忘记门偏差,来建立明确的多时间尺度语言模型。时间尺度的分布选择通过指数衰减的记忆细胞组合来接近自然语言的权力法统计。然后我们用词缩放实验和忘记门可视化来对通过模型每个部分所传递的信息的时间尺度进行实证分析。这些实验表明,多时间尺度模型成功地在理想的时间尺度内学习了演示,而且分布包括比标准的 LSTM 更长的时间尺度。此外,关于高、中、低频词汇的信息会以适当时间尺度的单位为优先选择。因此,我们展示了如何构建语言模型,对不同的信息时间尺度进行解释表达。