Recommender system practitioners are facing increasing pressure to explain recommendations. We explore how to explain recommendations using counterfactual logic, i.e. "Had you not interacted with the following items before, it is likely we would not recommend this item." Compared to traditional explanation logic, counterfactual explanations are easier to understand and more technically verifiable. The major challenge of generating such explanations is the computational cost because it requires repeatedly retraining the models to obtain the effect on a recommendation caused by removing user (interaction) history. We propose a learning-based framework to generate counterfactual explanations. The key idea is to train a surrogate model to learn the effect of removing a subset of user history on the recommendation. To this end, we first artificially simulate the counterfactual outcomes on the recommendation after deleting subsets of history. Then we train surrogate models to learn the mapping between a history deletion and the change in the recommendation caused by the deletion. Finally, to generate an explanation, we find the history subset predicted by the surrogate model that is most likely to remove the recommendation. Through offline experiments and online user studies, we show our method, compared to baselines, can generate explanations that are more counterfactually valid and more satisfactory considered by users.
翻译:建议系统从业者面临越来越大的压力来解释建议。我们探索如何用反事实逻辑来解释建议,即“如果你以前没有与以下项目互动,我们很可能不会推荐这个项目。”与传统解释逻辑相比,反事实解释更容易理解,在技术上可以比较容易核查。产生这种解释的主要挑战是计算成本,因为它要求反复再培训模型,以获得对删除用户(互动)历史的建议的影响。我们建议了一个基于学习的框架,以产生反事实解释。关键的想法是培训一个替代模型,以学习删除建议中用户历史的一部分的效果。为此,我们首先在删除历史的子集后人为模拟建议中的反事实结果。然后我们培训替代模型,以学习历史删除和删除后的建议修改之间的映射。最后,为了产生解释,我们发现一个最有可能删除建议的代孕模型所预测的历史子。通过离线实验和在线用户研究,我们通过比照基线来展示我们的方法,比基准用户更令人满意地产生相反的解释。