Negative sampling, which samples negative triplets from non-observed ones in knowledge graph (KG), is an essential step in KG embedding. Recently, generative adversarial network (GAN), has been introduced in negative sampling. By sampling negative triplets with large gradients, these methods avoid the problem of vanishing gradient and thus obtain better performance. However, they make the original model more complex and harder to train. In this paper, motivated by the observation that negative triplets with large gradients are important but rare, we propose to directly keep track of them with the cache. In this way, our method acts as a "distilled" version of previous GAN-based methods, which does not waste training time on additional parameters to fit the full distribution of negative triplets. However, how to sample from and update the cache are two critical questions. We propose to solve these issues by automated machine learning techniques. The automated version also covers GAN-based methods as special cases. Theoretical explanation of NSCaching is also provided, justifying the superior over fixed sampling scheme. Besides, we further extend NSCaching with skip-gram model for graph embedding. Finally, extensive experiments show that our method can gain significant improvements on various KG embedding models and the skip-gram model, and outperforms the state-of-the-art negative sampling methods.
翻译:在知识图(KG)中,从未观测到的知识图(KG)中进行负三重抽样,这是KG嵌入的一个必要步骤。最近,在负面抽样中引入了基因对抗网络(GAN),通过用大梯度对负三重抽样进行抽样,这些方法避免了梯度消失的问题,从而取得更好的性能。然而,它们使原始模型更加复杂和难于培训。在本文中,由于观察到具有大梯度的负三重抽样是重要的但很少,我们提议直接用缓存来跟踪它们。这样,我们的方法就成了以前基于GAN的方法的“留存”版本,它不会浪费额外参数的培训时间,以适应负三重梯度的全面分布。然而,如何从缓冲中取样并更新是两个关键问题。我们提议通过自动化机器学习技术解决这些问题。自动化版本还包含以GAN为基础的方法,作为特殊案例。对NSCaching的理论性解释也提供了对固定取样方法的优越性。此外,我们进一步扩展了NSCaching(VG)模型的“留存”版本,用跳式模型来展示各种图表模型。最后,可以展示我们的重要模型的模型,并展示各种模型。