Fairness and Outlier Detection (OD) are closely related, as it is exactly the goal of OD to spot rare, minority samples in a given population. However, when being a minority (as defined by protected variables, such as race/ethnicity/sex/age) does not reflect positive-class membership (such as criminal/fraud), OD produces unjust outcomes. Surprisingly, fairness-aware OD has been almost untouched in prior work, as fair machine learning literature mainly focuses on supervised settings. Our work aims to bridge this gap. Specifically, we develop desiderata capturing well-motivated fairness criteria for OD, and systematically formalize the fair OD problem. Further, guided by our desiderata, we propose FairOD, a fairness-aware outlier detector that has the following desirable properties: FairOD (1) exhibits treatment parity at test time, (2) aims to flag equal proportions of samples from all groups (i.e. obtain group fairness, via statistical parity), and (3) strives to flag truly high-risk samples within each group. Extensive experiments on a diverse set of synthetic and real world datasets show that FairOD produces outcomes that are fair with respect to protected variables, while performing comparable to (and in some cases, even better than) fairness-agnostic detectors in terms of detection performance.
翻译:公平与外部探测(OD)密切相关,因为它恰恰是OD的目标,目的是在特定人群中发现稀有的、少数的样本;然而,当少数群体(按照种族/族裔/性别/年龄等受保护变量的定义)不能反映正面阶级成员(如犯罪/欺诈)时,OD会产生不公正的结果;令人惊讶的是,公平意识的OD几乎没有出现在先前的工作中,因为公平的机器学习文献主要侧重于受监督的环境;我们的工作旨在弥合这一差距。具体地说,我们为OD制定了具有良好动机的公平标准,并系统地将公平的OD问题正式化。此外,我们建议公平认识的外部探测器,具有以下适当特性:FairOD(1) 测试时的对待对等,(2) 目的是将所有群体的样本比例(即通过统计均等获得群体公平性)与(3) 努力在每一个群体中标出真正高风险的样本。关于多种合成和真实的世界数据集的广泛实验表明,在我们的 desilatralator 中,公平性的结果甚至可以比较性地检测性,同时进行有保护的变量。