Most existing approaches for class expression learning in description logics are search algorithms. As the search space of these approaches is infinite, they often fail to scale to large learning problems. Our main intuition is that class expression learning can be regarded as a translation problem. Based thereupon, we propose a new family of class expression learning approaches which we dub neural class expression synthesis. Instances of this new family circumvent the high search costs entailed by current algorithms by translating training examples into class expressions in a fashion akin to machine translation solutions. Consequently, they are not subject to the runtime limitations of search-based approaches post training. We study three instances of this novel family of approaches to synthesize class expressions from sets of positive and negative examples. An evaluation of our approach on four benchmark datasets suggests that it can effectively synthesize high-quality class expressions with respect to the input examples in approximately one second on average. Moreover, a comparison to other state-of-the-art approaches suggests that we achieve better F-measures on large datasets. For reproducibility purposes, we provide our implementation as well as pretrained models in our public GitHub repository at https://github.com/fosterreproducibleresearch/NCES
翻译:由于这些方法的搜索空间是无限的,因此往往不能扩大到大型学习问题。我们的主要直觉是,课堂表达学习可被视为翻译问题。基于此,我们提出一个新的班级表达学习方法系列,我们据此对神经类表达表达进行合成。这个新家庭的例子通过将培训实例转换成类似于机器翻译解决方案的时装,绕过当前算法带来的高搜索成本。因此,这些方法的搜索空间往往没有局限在搜索方法后培训的运行时间范围内。我们研究了三个实例,即从一系列正面和负面的例子中合成班级表达方式的新颖方法。我们对四个基准数据集的方法的评估表明,它能够有效地将高质量的班级表达方式与输入实例平均大约一秒钟中的数字结合起来。此外,与其他最先进的算法方法的比较表明,我们在大型数据集上实现了更好的F度测量方法。为了再生化目的,我们提供了我们的实施方法以及我们公共GitHubsresprespreubus/fofcom公共Gestregistrate模型。