Learning from implicit feedback is challenging because of the difficult nature of the one-class problem: we can observe only positive examples. Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem. However, such methods have two main drawbacks particularly in large-scale applications; (1) the pairwise approach is severely inefficient due to the quadratic computational cost; and (2) even recent model-based samplers (e.g. IRGAN) cannot achieve practical efficiency due to the training of an extra model. In this paper, we propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart while performing similarly to the pairwise counterpart in terms of ranking effectiveness. Our approach estimates the probability densities of positive items for each user within a rich class of distributions, viz. \emph{exponential family}. In our formulation, we derive a loss function and the appropriate negative sampling distribution based on maximum likelihood estimation. We also develop a practical technique for risk approximation and a regularisation scheme. We then discuss that our single-model approach is equivalent to an IRGAN variant under a certain condition. Through experiments on real-world datasets, our approach outperforms the pointwise and pairwise counterparts in terms of effectiveness and efficiency.
翻译:从隐含的反馈中学习具有挑战性,因为单级问题具有困难的性质:我们只能看到正面的例子。大多数常规方法都采用对等排名办法和负抽样来应对单级问题。但是,这类方法有两个主要缺点,特别是在大规模应用方面;(1) 双向方法由于二次计算成本而严重无效;以及(2) 即使是最近基于模型的采样者(例如IRGAN)也由于培训额外模型而无法实现实际效率。在本文中,我们建议采用一种从头到尾的学习方法,这种方法达到与点对口相似的趋同速度,同时在等级效力方面与对口的对口方法相似。我们的方法估计了在丰富的分配类别(即 \emph{Exponitial family})中每个用户正项的概率密度。在我们的提法中,我们从最大可能性估计中得出损失的功能和适当的负抽样分布。我们还在风险近似值和正规化计划方面开发了一种实用技术。我们随后讨论的单一模型方法相当于IRGAN方法在某种条件下对准效率的变异。