Current breakthroughs in natural language processing have benefited dramatically from neural language models, through which distributional semantics can leverage neural data representations to facilitate downstream applications. Since neural embeddings use context prediction on word co-occurrences to yield dense vectors, they are inevitably prone to capture more semantic association than semantic similarity. To improve vector space models in deriving semantic similarity, we post-process neural word embeddings through deep metric learning, through which we can inject lexical-semantic relations, including syn/antonymy and hypo/hypernymy, into a distributional space. We introduce hierarchy-fitting, a novel semantic specialization approach to modelling semantic similarity nuances inherently stored in the IS-A hierarchies. Hierarchy-fitting attains state-of-the-art results on the common- and rare-word benchmark datasets for deriving semantic similarity from neural word embeddings. It also incorporates an asymmetric distance function to specialize hypernymy's directionality explicitly, through which it significantly improves vanilla embeddings in multiple evaluation tasks of detecting hypernymy and directionality without negative impacts on semantic similarity judgement. The results demonstrate the efficacy of hierarchy-fitting in specializing neural embeddings with semantic relations in late fusion, potentially expanding its applicability to aggregating heterogeneous data and various knowledge resources for learning multimodal semantic spaces.
翻译:自然语言处理的当前突破得益于神经语言模型,通过这些模型,分布式语义可以将神经数据表示方式用于促进下游应用。由于神经内嵌利用对单词共发量的背景预测来生成密度矢量,因此它们不可避免地会比语义相似性更容易捕捉更多的语义联系。为了改进矢量空间模型,得出语义相似性,我们通过深度的学习,将流程后神经字嵌入神经字嵌入。通过这些模型,我们可将词义-语义关系,包括同源/异名和机能/机能性,引入一个分配空间。由于神经内嵌入将使用对单词共发量数据进行环境预测,因此神经内嵌使用新颖的语义化专业化方法来模拟在IS-A等级结构中固有的语义相似性差异性,因此它们不可避免地能够捕捉到与稀有和稀有的语义基准数据集的状态,以便从神经文义嵌入的语义相似性词义关系中,我们还包含一个不对称的远程功能,以特别地将超重尼基的定向性明确化。通过这个方法,极大地改进了在高统制的内基化的内嵌化结构关系中,从而显示其潜在的判断性层次结构内脏内存的多重判断结果。