Keeping the performance of language technologies optimal as time passes is of great practical interest. Here we survey prior work concerned with the effect of time on system performance, establishing more nuanced terminology for discussing the topic and proper experimental design to support solid conclusions about the observed phenomena. We present a set of experiments with systems powered by large neural pretrained representations for English to demonstrate that {\em temporal model deterioration} is not as big a concern, with some models in fact improving when tested on data drawn from a later time period. It is however the case that {\em temporal domain adaptation} is beneficial, with better performance for a given time period possible when the system is trained on temporally more recent data. Our experiments reveal that the distinctions between temporal model deterioration and temporal domain adaptation becomes salient for systems built upon pretrained representations. Finally we examine the efficacy of two approaches for temporal domain adaptation without human annotations on new data, with self-labeling proving to be superior to continual pre-training. Notably, for named entity recognition, self-labeling leads to better temporal adaptation than human annotation.
翻译:当时间流逝时,保持语言技术的最佳性能具有极大的实际意义。在这里,我们调查以前与时间对系统性能的影响有关的工作,为讨论这个专题建立更细微的术语,并进行适当的实验设计,以支持关于观察到的现象的可靠结论。我们展示了一套实验系统,该系统的动力是大型神经预设的英语演示,以证明“时间模型衰减”不是那么大的问题,有些模型在测试后一段时间的数据时实际上有所改进。然而,情况是,“时间域适应”是有好处的,当系统接受时间上更近的数据培训时,在一定的时期内,业绩可能更好。我们的实验表明,时间模型衰减和时间域适应之间的区别在预先测试后的系统中变得突出。最后,我们研究了在没有人类对新数据进行说明的情况下进行时间域适应的两种方法的功效,自我标记证明优于持续的培训前期。值得注意的是,对于被命名的实体的识别,自我标记导致时间适应优于人类说明。