A variety of contextualised language models have been proposed in the NLP community, which are trained on diverse corpora to produce numerous Neural Language Models (NLMs). However, different NLMs have reported different levels of performances in downstream NLP applications when used as text representations. We propose a sentence-level meta-embedding learning method that takes independently trained contextualised word embedding models and learns a sentence embedding that preserves the complementary strengths of the input source NLMs. Our proposed method is unsupervised and is not tied to a particular downstream task, which makes the learnt meta-embeddings in principle applicable to different tasks that require sentence representations. Specifically, we first project the token-level embeddings obtained by the individual NLMs and learn attention weights that indicate the contributions of source embeddings towards their token-level meta-embeddings. Next, we apply mean and max pooling to produce sentence-level meta-embeddings from token-level meta-embeddings. Experimental results on semantic textual similarity benchmarks show that our proposed unsupervised sentence-level meta-embedding method outperforms previously proposed sentence-level meta-embedding methods as well as a supervised baseline.
翻译:国家语言方案社区提出了各种背景语言模型,这些模型经过培训,以产生多种神经语言模型(NLMS),在不同的组合中产生多种神经语言模型(NLMS),但不同的NLMS报告了下游NLP应用程序在用作文本表述时的不同性能水平。我们提出了一个句级混合学习方法,该方法采用独立培训背景语言嵌入模型,并学习一个内嵌的句子,保留输入源NLMS的互补优势。我们建议的方法不受监督,而且不与特定的下游任务挂钩,这使得所学的元组成原则上适用于需要判刑表述的不同任务。具体地说,我们首先预测单个NLMS公司获得的象征性级嵌入,并学习关注度,表明源嵌入到其象征性水平元集成的源的贡献。接下来,我们采用平均值和最大程度的集合,从象征性的元组装组装中产生句级的元组装。关于语系文本的实验结果显示,我们提议的未加固的元模级嵌制前的元模制的元模集制方法显示我们提议的未经修正的元模制的元模制的元模制的元模制的元模制的元模方法。