As Artificial Intelligence (AI) is used in more applications, the need to consider and mitigate biases from the learned models has followed. Most works in developing fair learning algorithms focus on the offline setting. However, in many real-world applications data comes in an online fashion and needs to be processed on the fly. Moreover, in practical application, there is a trade-off between accuracy and fairness that needs to be accounted for, but current methods often have multiple hyperparameters with non-trivial interaction to achieve fairness. In this paper, we propose a flexible ensemble algorithm for fair decision-making in the more challenging context of evolving online settings. This algorithm, called FARF (Fair and Adaptive Random Forests), is based on using online component classifiers and updating them according to the current distribution, that also accounts for fairness and a single hyperparameters that alters fairness-accuracy balance. Experiments on real-world discriminated data streams demonstrate the utility of FARF.
翻译:由于人工智能(AI)被用于更多的应用中,因此需要考虑和减少从所学模型中产生的偏差。大多数开发公平学习算法的工作都侧重于离线设置。然而,在许多真实世界应用数据中,数据以在线方式出现,需要通过自动处理。此外,在实际应用中,准确性和公平性之间有一个权衡,需要加以核算,但目前的方法往往具有多种超参数,具有非三边互动,以实现公平。在本文中,我们提议在不断变化的在线设置更具挑战性的背景下,为公平决策提供一个灵活的共通算法。这个称为FARF(公平与适应随机森林)的算法基于使用在线组件分类器并根据当前分布进行更新,这也考虑到公平性和单一的超参数,从而改变公平与准确性平衡。对真实世界受歧视的数据流的实验证明了FARF的效用。