Concept learning approaches based on refinement operators explore partially ordered solution spaces to compute concepts, which are used as binary classification models for individuals. However, the number of concepts explored by these approaches can easily grow to the millions for complex learning problems. This often leads to impractical runtimes. We propose to alleviate this problem by predicting the length of target concepts before the exploration of the solution space. By these means, we can prune the search space during concept learning. To achieve this goal, we compare four neural architectures and evaluate them on four benchmarks. Our evaluation results suggest that recurrent neural network architectures perform best at concept length prediction with a macro F-measure ranging from 38% to 92%. We then extend the CELOE algorithm--which learns ALC concepts--with our concept length predictor. Our extension yields the algorithm CLIP. In our experiments, CLIP is at least 7.5 times faster than other state-of-the-art concept learning algorithms for ALC--including CELOE--and achieves significant improvements in the F-measure of the concepts learned on 3 out of 4 datasets. For reproducibility, we provide our implementation in the public GitHub repository at https://github.com/ConceptLengthLearner/LearnLengths
翻译:基于完善操作者的概念学习方法基于完善操作者探索部分有序的计算概念的解决方案空间,这些概念被用作个人的二元分类模型。然而,这些方法所探讨的概念数量很容易增长到数百万个复杂学习问题,这往往会导致不切实际的运行时间。我们提议通过在探索解决方案空间之前预测目标概念的长度来缓解这一问题。通过这些手段,我们可以在概念学习过程中利用搜索空间。为了实现这一目标,我们比较了四个神经结构,并根据四个基准对它们进行评估。我们的评价结果表明,经常神经网络结构在概念长度预测方面效果最好,其宏观F措施从38%到92%不等。我们随后扩展了CELOE算法,该算法用我们的概念长度预测器学习ALC概念。我们的扩展产生了CLIP算法。在我们的实验中,CLIP比其他州-包括CLOE-CLOE-和在4个数据集外的3个概念的F-测量方法有显著改进。我们提供了在MACUB/Lcommev/Gicom中执行我们的公共数据库。