Fairness and robustness are critical elements of Trustworthy AI that need to be addressed together. Fairness is about learning an unbiased model while robustness is about learning from corrupted data, and it is known that addressing only one of them may have an adverse affect on the other. In this work, we propose a sample selection-based algorithm for fair and robust training. To this end, we formulate a combinatorial optimization problem for the unbiased selection of samples in the presence of data corruption. Observing that solving this optimization problem is strongly NP-hard, we propose a greedy algorithm that is efficient and effective in practice. Experiments show that our algorithm obtains fairness and robustness that are better than or comparable to the state-of-the-art technique, both on synthetic and benchmark real datasets. Moreover, unlike other fair and robust training baselines, our algorithm can be used by only modifying the sampling step in batch selection without changing the training algorithm or leveraging additional clean data.
翻译:公平性和稳健性是值得信赖的AI的关键要素,需要一起解决。公平性是学习一个公正的模型,而稳健性则是从腐败的数据中学习,众所周知,只处理其中之一可能会对另一个数据产生不利影响。在这项工作中,我们建议为公平和稳健的培训制定基于抽样的筛选算法。为此,我们为在存在数据腐败的情况下不偏袒地选择样本制定了组合优化问题。注意到解决这一优化问题非常困难,我们提出了一种在实践上高效和高效的贪婪算法。实验表明,我们的算法在合成和基准真实数据集中都获得了比最新技术更好或可比的公正和稳健性。此外,与其他公平和稳健的培训基线不同,我们的算法只能通过修改批量选择中的抽样步骤而不改变培训算法或利用额外的清洁数据来使用。