Knowledge graph embedding (KGE) models are an effective and popular approach to represent and reason with multi-relational data. Prior studies have shown that KGE models are sensitive to hyperparameter settings, however, and that suitable choices are dataset-dependent. In this paper, we explore hyperparameter optimization (HPO) for very large knowledge graphs, where the cost of evaluating individual hyperparameter configurations is excessive. Prior studies often avoided this cost by using various heuristics; e.g., by training on a subgraph or by using fewer epochs. We systematically discuss and evaluate the quality and cost savings of such heuristics and other low-cost approximation techniques. Based on our findings, we introduce GraSH, an efficient multi-fidelity HPO algorithm for large-scale KGEs that combines both graph and epoch reduction techniques and runs in multiple rounds of increasing fidelities. We conducted an experimental study and found that GraSH obtains state-of-the-art results on large graphs at a low cost (three complete training runs in total).
翻译:知识图嵌入模型(KGE)是一种有效、流行的方法,它代表了多关系数据并说明了理由。但先前的研究显示,KGE模型对超参数设置十分敏感,适当的选择取决于数据集。在本文中,我们探索了对非常大的知识图的超参数优化(HPO ), 评估个人超参数配置的成本过高。先前的研究往往通过使用各种超理论学来避免这一成本;例如,通过子图培训或使用较少的粒子。我们系统地讨论和评估了这种超光谱和其他低成本近似技术的质量和成本节约。我们根据我们的调查结果,为大型知识图组引入了高效的多纤维速率 HPO算法,该算法结合了图形和粒子减少技术,并且以多种周期的忠心运行。我们进行了一项实验研究,发现GRASH以低成本获得了大图上的最新结果(总共进行了三次完整的培训 )。