Word embeddings are extensively used in various NLP problems as a state-of-the-art semantic feature vector representation. Despite their success on various tasks and domains, they might exhibit an undesired bias for stereotypical categories due to statistical and societal biases that exist in the dataset they are trained on. In this study, we analyze the gender bias in four different pre-trained word embeddings specifically for the depression category in the mental disorder domain. We use contextual and non-contextual embeddings that are trained on domain-independent as well as clinical domain-specific data. We observe that embeddings carry bias for depression towards different gender groups depending on the type of embeddings. Moreover, we demonstrate that these undesired correlations are transferred to the downstream task for depression phenotype recognition. We find that data augmentation by simply swapping gender words mitigates the bias significantly in the downstream task.
翻译:语言嵌入被广泛用于各种国家语言方案问题,作为一种最新的语义特征矢量代表。尽管它们在不同任务和领域取得了成功,但由于它们所培训的数据集中存在的统计和社会偏见,它们可能对陈规定型类别表现出不理想的偏见。在这项研究中,我们分析了四种专门为精神失常领域的抑郁类别而预先培训的单词嵌入中的性别偏见。我们使用在领域独立和临床特定领域数据方面受过培训的背景和非文字嵌入。我们发现,根据嵌入类型,嵌入会给不同性别群体带来抑郁的偏见。此外,我们证明这些不理想的关联性被转移到下游的抑郁苯型识别任务中。我们发现,仅仅通过性别词互换就可以使下游任务中的偏见大大减轻。