Learning-to-rank (LTR) has become a key technology in E-commerce applications. Most existing LTR approaches follow a supervised learning paradigm from offline labeled data collected from the online system. However, it has been noticed that previous LTR models can have a good validation performance over offline validation data but have a poor online performance, and vice versa, which implies a possible large inconsistency between the offline and online evaluation. We investigate and confirm in this paper that such inconsistency exists and can have a significant impact on AliExpress Search. Reasons for the inconsistency include the ignorance of item context during the learning, and the offline data set is insufficient for learning the context. Therefore, this paper proposes an evaluator-generator framework for LTR with item context. The framework consists of an evaluator that generalizes to evaluate recommendations involving the context, and a generator that maximizes the evaluator score by reinforcement learning, and a discriminator that ensures the generalization of the evaluator. Extensive experiments in simulation environments and AliExpress Search online system show that, firstly, the classic data-based metrics on the offline dataset can show significant inconsistency with online performance, and can even be misleading. Secondly, the proposed evaluator score is significantly more consistent with the online performance than common ranking metrics. Finally, as the consequence, our method achieves a significant improvement (\textgreater$2\%$) in terms of Conversion Rate (CR) over the industrial-level fine-tuned model in online A/B tests.
翻译:现有LTR方法大多遵循从在线系统收集的离线标签数据提供的受监督的学习模式,然而,人们注意到,以前的LTR模型可以在离线验证数据上有一个良好的验证性业绩,但在线绩效较差,反之亦然,这意味着离线和在线评价之间可能存在很大的不一致。我们在本文件中调查和确认,这种不一致存在,并可能对AliExpress搜索产生重大影响。不一致的原因包括学习过程中对项目背景的无知,而离线数据集不足以学习背景。因此,本文件提议为LTR项目背景的LTR建立一个评价者-生成框架。框架包括一名评价者,负责对涉及背景的建议进行一般性评价,以及一个通过强化学习使评价者得分最大化的生成者,以及一个确保评价者普遍化的区分者。在模拟环境和AliExpress搜索在线系统中的广泛实验表明,在离线数据集中基于数据的典型指标首先可以显示,在在线业绩测试方面与在线评级相比明显不相符,最后,在标准评级方面,在在线评级方面,比共同的评级更具有误导性。