The growing capability and availability of generative language models has enabled a wide range of new downstream tasks. Academic research has identified, quantified and mitigated biases present in language models but is rarely tailored to downstream tasks where wider impact on individuals and society can be felt. In this work, we leverage one popular generative language model, GPT-3, with the goal of writing unbiased and realistic job advertisements. We first assess the bias and realism of zero-shot generated advertisements and compare them to real-world advertisements. We then evaluate prompt-engineering and fine-tuning as debiasing methods. We find that prompt-engineering with diversity-encouraging prompts gives no significant improvement to bias, nor realism. Conversely, fine-tuning, especially on unbiased real advertisements, can improve realism and reduce bias.
翻译:学术研究发现、量化和减轻了语言模式中存在的偏见,但很少适应对个人和社会产生更广泛影响的下游任务;在这项工作中,我们利用一种流行的基因化语言模式GPT-3,目的是编写公正和现实的招工广告;我们首先评估零弹广告的偏向和现实主义,并将其与现实世界广告进行比较;然后我们评价即时工程和微调,将其作为贬低偏见的方法;我们发现,以鼓励多样性的速效来迅速工程,对偏见和现实主义没有重大改进。相反,微调,特别是对无偏见真实广告的微调,可以改善现实主义,减少偏见。