Recommender system practitioners are facing increasing pressure to explain recommendations. We explore how to explain recommendations using counterfactual logic, i.e. "Had you not interacted with the following items, we would not recommend it." Compared to the traditional explanation logic, counterfactual explanations are easier to understand, more technically verifiable, and more informative in terms of giving users control over recommendations. The major challenge of generating such explanations is the computational cost because it requires repeatedly retraining the models to obtain the effect on a recommendation caused by the absence of user history. We propose a learning-based framework to generate counterfactual explanations. The key idea is to train a surrogate model to learn the effect of removing a subset of user history on the recommendation. To this end, we first artificially simulate the counterfactual outcomes on the recommendation after deleting subsets of history. Then we train a surrogate model to learn the mapping between a history deletion and the corresponding change of the recommendation caused by the deletion. Finally, to generate an explanation, we find the history subset predicted by the surrogate model that is most likely to remove the recommendation. Through offline experiments and online user studies, we show our method, compared to baselines, can generate explanations that are more counterfactually valid and more satisfactory considered by users.
翻译:建议系统从业者面临越来越大的压力来解释建议。我们探索如何用反事实逻辑来解释建议,即“如果你没有与以下项目互动,我们不会推荐它。”与传统的解释逻辑相比,反事实解释更容易理解,在让用户控制建议方面,比较技术上可以核查,在让建议得到更多信息方面,反事实解释比较容易理解,比较技术上可以核查,比较信息丰富。产生这种解释的主要挑战是计算成本,因为它要求反复对模型进行再培训,以获得对因用户历史缺失而导致的建议的影响。我们提出了一个基于学习的框架,以产生反事实解释。关键的想法是训练一个替代模型,以学习删除建议中用户历史的一部分的影响。为此,我们首先在删除历史的子集之后,人为模拟建议中的反事实结果。然后,我们训练一个替代模型,以学习删除历史和删除建议的相应修改。最后,我们发现由替代模型所预测的历史子集最有可能删除建议。通过离线实验和在线用户研究,我们通过对比比较基线,可以比较我们的方法是更准确的。