In this paper, we explore the task of automatically generating natural language descriptions of salient patterns in a time series, such as stock prices of a company over a week. A model for this task should be able to extract high-level patterns such as presence of a peak or a dip. While typical contemporary neural models with attention mechanisms can generate fluent output descriptions for this task, they often generate factually incorrect descriptions. We propose a computational model with a truth-conditional architecture which first runs small learned programs on the input time series, then identifies the programs/patterns which hold true for the given input, and finally conditions on only the chosen valid program (rather than the input time series) to generate the output text description. A program in our model is constructed from modules, which are small neural networks that are designed to capture numerical patterns and temporal information. The modules are shared across multiple programs, enabling compositionality as well as efficient learning of module parameters. The modules, as well as the composition of the modules, are unobserved in data, and we learn them in an end-to-end fashion with the only training signal coming from the accompanying natural language text descriptions. We find that the proposed model is able to generate high-precision captions even though we consider a small and simple space of module types.
翻译:在本文中, 我们探索了在时间序列中自动生成突出模式的自然语言描述的任务, 如公司一周的股票价格。 任务模型应该能够提取高层次模式, 如峰值或底值的存在。 典型的当代神经模型和关注机制可以产生流畅的输出描述, 它们往往产生不真实的描述。 我们提出了一个计算模型, 包含一个真实条件的架构, 它首先在输入时间序列上运行小的学习程序, 然后确定对给定输入真实的程序/ 模式, 并且只有选择的有效程序( 而不是输入时间序列) 才能最终产生输出文本描述。 我们模型中的一个程序来自模块, 是小型的神经网络, 设计来捕捉数字模式和时间信息。 这些模块在多个程序中共享, 能够组成以及高效地学习模块参数。 模块以及模块的构成在数据中是无法观察到的, 我们以最终到它们的方式学习它们, 唯一的培训信号来自随附的自然语言版图示的模块。 我们发现, 高的模型是能够生成的。