It is fair to say that many of the prominent examples of bias in Machine Learning (ML) arise from bias that is there in the training data. In fact, some would argue that supervised ML algorithms cannot be biased, they reflect the data on which they are trained. In this paper we demonstrate how ML algorithms can misrepresent the training data through underestimation. We show how irreducible error, regularization and feature and class imbalance can contribute to this underestimation. The paper concludes with a demonstration of how the careful management of synthetic counterfactuals can ameliorate the impact of this underestimation bias.
翻译:可以公平地说,机器学习中许多明显的偏见例子都源于培训数据中存在的偏见。事实上,有些人会争辩说,受监督的ML算法不能有偏见,它们反映的是培训所依据的数据。在本文中,我们展示了ML算法如何通过低估来歪曲培训数据。我们显示了不可避免的错误、正规化、特征和阶级不平衡是如何促成这种低估的。文件最后展示了对合成反事实的审慎管理如何能够减轻这种低估偏见的影响。