Contextualized embeddings such as BERT can serve as strong input representations to NLP tasks, outperforming their static embeddings counterparts such as skip-gram, CBOW and GloVe. However, such embeddings are dynamic, calculated according to a sentence-level context, which limits their use in lexical semantics tasks. We address this issue by making use of dynamic embeddings as word representations in training static embeddings, thereby leveraging their strong representation power for disambiguating context information. Results show that this method leads to improvements over traditional static embeddings on a range of lexical semantics tasks, obtaining the best reported results on seven datasets.
翻译:BERT等背景嵌入器可成为NLP任务的有力输入代表,优于其静态嵌入器,如跳格、CBOW和GloVe。然而,这种嵌入器是动态的,根据判决级别计算,限制其在词汇语义任务中的使用。我们通过在培训静态嵌入器中使用动态嵌入器作为文字表达器来解决这一问题,从而利用强大的代表力来模糊背景信息。结果显示,这种方法可以改进传统静态嵌入系统在一系列词汇语义任务上的功能,获得七个数据集的最佳报告结果。