As Learning-to-Rank (LTR) approaches primarily seek to improve ranking quality, their output scores are not scale-calibrated by design -- for example, adding a constant to the score of each item on the list will not affect the list ordering. This fundamentally limits LTR usage in score-sensitive applications. Though a simple multi-objective approach that combines a regression and a ranking objective can effectively learn scale-calibrated scores, we argue that the two objectives can be inherently conflicting, which makes the trade-off far from ideal for both of them. In this paper, we propose a novel regression compatible ranking (RCR) approach to achieve a better trade-off. The advantage of the proposed approach is that the regression and ranking components are well aligned which brings new opportunities for harmonious regression and ranking. Theoretically, we show that the two components share the same minimizer at global minima while the regression component ensures scale calibration. Empirically, we show that the proposed approach performs well on both regression and ranking metrics on several public LTR datasets, and significantly improves the Pareto frontiers in the context of multi-objective optimization. Furthermore, we evaluated the proposed approach on YouTube Search and found that it not only improved the ranking quality of the production pCTR model, but also brought gains to the click prediction accuracy.
翻译:由于 " 学习到兰克(LTR) " 方法主要力求提高排名质量,它们的产出分数没有按设计进行比例校准 -- -- 例如,在列表中每个项目的得分上增加一个常数不会影响列表顺序。这从根本上限制了在对分敏感的应用程序中使用LTR。虽然一种简单的多目标方法,结合回归和排序目标,可以有效地学习比例校准分数,但我们认为,这两个目标本身可能相互冲突,使两者的取舍远非理想。在本文中,我们建议采用新的回归兼容排序(RCR)方法,以便实现更好的权衡。拟议方法的优点是,回归和排序各组成部分完全一致,为和谐回归和排序带来新的机会。理论上,我们表明,两个组成部分在全球迷你中共享相同的最小值,而回归部分则确保比例校准。我们很生动地指出,拟议的方法在回归和对两个公共LTR数据集的评分都远远达不到理想。我们建议采用新的回归和排序方法,在多目标的精确度范围内大大改进了Preto边界。此外,我们还评估了在多目标的精确度预测中改进了PTR的质量。我们还评估了它。