Machine learning models built on datasets containing discriminative instances attributed to various underlying factors result in biased and unfair outcomes. It's a well founded and intuitive fact that existing bias mitigation strategies often sacrifice accuracy in order to ensure fairness. But when AI engine's prediction is used for decision making which reflects on revenue or operational efficiency such as credit risk modelling, it would be desirable by the business if accuracy can be somehow reasonably preserved. This conflicting requirement of maintaining accuracy and fairness in AI motivates our research. In this paper, we propose a fresh approach for simultaneous improvement of fairness and accuracy of ML models within a realistic paradigm. The essence of our work is a data preprocessing technique that can detect instances ascribing a specific kind of bias that should be removed from the dataset before training and we further show that such instance removal will have no adverse impact on model accuracy. In particular, we claim that in the problem settings where instances exist with similar feature but different labels caused by variation in protected attributes , an inherent bias gets induced in the dataset, which can be identified and mitigated through our novel scheme. Our experimental evaluation on two open-source datasets demonstrates how the proposed method can mitigate bias along with improving rather than degrading accuracy, while offering certain set of control for end user.
翻译:在包含各种基本因素导致偏向和不公平结果的歧视性实例的数据集上建立的机床学习模型含有各种基本因素的歧视性实例,这是一个有充分依据和直观的事实是,现有的减少偏向战略往往牺牲准确性以确保公平性。但是,当AI引擎的预测用于反映收入或业务效率的决策时,如信用风险模型等反映收入或业务效率时,商业界最好能够以某种合理的方式保持准确性。在AI中保持准确性和公正性的这一相互矛盾的要求促使我们进行研究。在本文件中,我们提出了一个在现实的范式内同时提高ML模型的公正和准确性的新办法。我们工作的精髓是一种数据处理技术,它能够检测在培训前从数据集中删除某种特定偏差的事例。我们进一步表明,这种删除不会对模型准确性产生不利影响。特别是,我们声称,在存在类似特征但因受保护属性的变化而造成不同标签的问题环境中,一个内在的偏差会引出数据集,可以通过我们的新方案加以识别和减轻。我们对两个开源数据集的实验性评价方法,可以说明如何在降低某些控制方面减少偏差,而同时提供某种偏差的偏差。