There are several measures for fairness in ranking, based on different underlying assumptions and perspectives. PL optimization with the REINFORCE algorithm can be used for optimizing black-box objective functions over permutations. In particular, it can be used for optimizing fairness measures. However, though effective for queries with a moderate number of repeating sessions, PL optimization has room for improvement for queries with a small number of repeating sessions. In this paper, we present a novel way of representing permutation distributions, based on the notion of permutation graphs. Similar to PL, our distribution representation, called PPG, can be used for black-box optimization of fairness. Different from PL, where pointwise logits are used as the distribution parameters, in PPG pairwise inversion probabilities together with a reference permutation construct the distribution. As such, the reference permutation can be set to the best sampled permutation regarding the objective function, making PPG suitable for both deterministic and stochastic rankings. Our experiments show that PPG, while comparable to PL for larger session repetitions (i.e., stochastic ranking), improves over PL for optimizing fairness metrics for queries with one session (i.e., deterministic ranking). Additionally, when accurate utility estimations are available, e.g., in tabular models, the performance of PPG in fairness optimization is significantly boosted compared to lower quality utility estimations from a learning to rank model, leading to a large performance gap with PL. Finally, the pairwise probabilities make it possible to impose pairwise constraints such as "item $d_1$ should always be ranked higher than item $d_2$." Such constraints can be used to simultaneously optimize the fairness metric and control another objective such as ranking performance.
翻译:根据不同的基本假设和观点,在排名上有一些公平度的衡量标准。 与REINFORCE 算法的PL优化可用于优化黑盒公平性功能, 从而优化黑盒目标功能, 特别是, 它可以优化公平度措施。 但是, 虽然对询问有效, 重复会话数量略少, PL优化有改进的余地。 在本文中, 我们展示了一种新颖的方式, 代表调整分布, 其依据是平整图的概念。 类似 PL, 我们的分布代表, 叫做 PPG, 可用于优化黑盒公平性能优化。 不同于 PL, 在 PL, 使用点对点对点对点对点的对点对点对点对点对点对点, 加上一个参考对点对点对点, 这样, 参照点对点对点, 使PGPG适合确定性价比值的排序。 我们的实验显示, PPPGG, 与 PLC 相比, 更大规模的重复( i. e., stochacrictal 排名) 的对点对点的对点,, 质对点对点对点, 质性能的排序, 质质质质质质质评估,,, 的排序, 与比对点对点, 当点, 当点, 对点, 当对点, 当点, 当点, 当点, 当点对点, 当点, 当点, 当点, 当点, 当点对点对点对点对点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点, 当点