Encoded text representations often capture sensitive attributes about individuals (e.g., race or gender), which raise privacy concerns and can make downstream models unfair to certain groups. In this work, we propose FEDERATE, an approach that combines ideas from differential privacy and adversarial training to learn private text representations which also induces fairer models. We empirically evaluate the trade-off between the privacy of the representations and the fairness and accuracy of the downstream model on four NLP datasets. Our results show that FEDERATE consistently improves upon previous methods, and thus suggest that privacy and fairness can positively reinforce each other.
翻译:在这项工作中,我们提议FEDERATE, 这种方法结合了从不同的隐私和对抗性培训到学习私人文本代表的想法,这又能产生更公平的模式。我们从经验上评估了在四个国家语言平台数据集中代表的隐私与下游模式的公平性和准确性之间的权衡。我们的结果表明,FEDERATE不断改进以往的方法,从而表明隐私和公平可以积极加强彼此。