Interpretable Learning to Rank (LtR) is an emerging field within the research area of explainable AI, aiming at developing intelligible and accurate predictive models. While most of the previous research efforts focus on creating post-hoc explanations, in this paper we investigate how to train effective and intrinsically-interpretable ranking models. Developing these models is particularly challenging and it also requires finding a trade-off between ranking quality and model complexity. State-of-the-art rankers, made of either large ensembles of trees or several neural layers, exploit in fact an unlimited number of feature interactions making them black boxes. Previous approaches on intrinsically-interpretable ranking models address this issue by avoiding interactions between features thus paying a significant performance drop with respect to full-complexity models. Conversely, ILMART, our novel and interpretable LtR solution based on LambdaMART, is able to train effective and intelligible models by exploiting a limited and controlled number of pairwise feature interactions. Exhaustive and reproducible experiments conducted on three publicly-available LtR datasets show that ILMART outperforms the current state-of-the-art solution for interpretable ranking of a large margin with a gain of nDCG of up to 8%.
翻译:在可解释的AI(LtR)研究领域,对排名进行解释性学习(LtR)是一个新兴领域,目的是开发可理解和准确的预测模型。虽然以前的大部分研究努力都侧重于创建热后解释,但我们在本文中调查如何培训有效和内在解释性排名模型。开发这些模型特别具有挑战性,还需要在排名质量和模型复杂性之间找到权衡。由大树群或多层神经层组成的最先进的排名者,事实上利用了无限数量的特征互动,形成黑盒。以前在内在解释性排名模型上采用的方法通过避免功能之间的互动来解决这一问题,从而在完全兼容性模型方面实现显著的绩效下降。反之,ILMART,我们基于LambdaMART的新颖和可解释的LtR解决方案,能够通过利用有限和受控数量的配对地特征互动来培训有效和不易理解的模式。在三个公开使用的LtR数据集上进行的深入和可复制的实验显示,ILARTMER 超越了当前8模型的可获取性差值。