The gender bias present in the data on which language models are pre-trained gets reflected in the systems that use these models. The model's intrinsic gender bias shows an outdated and unequal view of women in our culture and encourages discrimination. Therefore, in order to establish more equitable systems and increase fairness, it is crucial to identify and mitigate the bias existing in these models. While there is a significant amount of work in this area in English, there is a dearth of research being done in other gendered and low resources languages, particularly the Indian languages. English is a non-gendered language, where it has genderless nouns. The methodologies for bias detection in English cannot be directly deployed in other gendered languages, where the syntax and semantics vary. In our paper, we measure gender bias associated with occupations in Hindi language models. Our major contributions in this paper are the construction of a novel corpus to evaluate occupational gender bias in Hindi, quantify this existing bias in these systems using a well-defined metric, and mitigate it by efficiently fine-tuning our model. Our results reflect that the bias is reduced post-introduction of our proposed mitigation techniques. Our codebase is available publicly.
翻译:在经过预先培训的语文模式的数据中存在的性别偏见反映在使用这些模式的系统中。该模式的内在性别偏见表明,在我们的文化中,妇女的观点已经过时和不平等,并鼓励歧视。因此,为了建立更公平的制度和增加公平性,至关重要的是要查明和减少这些模式中存在的偏见。虽然在英语方面有大量的工作,但在其他性别语言上和低资源语言上,特别是印度语言上却缺乏研究。英语是一种非性别语言,没有性别名。英语是非性别语言,没有性别名。英语的偏见检测方法不能直接用于其他性别语言中,因为语法和语法各不相同。我们在我们的文件中,衡量与印地语模式中职业相关的性别偏见。我们在本文中的主要贡献是制作了一套新书,用以评价印地语中的职业性别偏见,用一个明确界定的尺度来量化这些系统中存在的这种偏见,并通过有效地微调我们的模型来减轻这种偏见。我们的结果表明,我们提议的减缓方法的偏见正在减少。我们的代码库是公开的。