Often, the data used to train ranking models is subject to label noise. For example, in web-search, labels created from clickstream data are noisy due to issues such as insufficient information in item descriptions on the SERP, query reformulation by the user, and erratic or unexpected user behavior. In practice, it is difficult to handle label noise without making strong assumptions about the label generation process. As a result, practitioners typically train their learning-to-rank (LtR) models directly on this noisy data without additional consideration of the label noise. Surprisingly, we often see strong performance from LtR models trained in this way. In this work, we describe a class of noise-tolerant LtR losses for which empirical risk minimization is a consistent procedure, even in the context of class-conditional label noise. We also develop noise-tolerant analogs of commonly used loss functions. The practical implications of our theoretical findings are further supported by experimental results.
翻译:通常情况下,用于培训排名模型的数据会受到标签噪音的影响,例如,在网上搜索中,通过点击流数据创建的标签由于以下问题而变得吵闹不休:诸如关于ERP的项目说明中信息不足、用户重新拟订查询、用户行为变化不定或出乎意料等;实际上,如果不对标签生成过程作出强有力的假设,就很难处理标签噪音问题;因此,从业人员通常不考虑标签噪音,直接用这种噪音数据来培训他们的学习到排序模型;令人惊讶的是,我们经常看到通过这种方式培训的LtR模型的强大性能。在这项工作中,我们描述了一种耐噪LtR损失类别,其经验风险最小化是一种一致的程序,即使在类装标签噪音的情况下也是如此。我们还开发了常用损失功能的耐噪模拟。我们理论发现的实际影响还得到了实验结果的进一步支持。