Recent advancements in GANs and diffusion models have enabled the creation of high-resolution, hyper-realistic images. However, these models may misrepresent certain social groups and present bias. Understanding bias in these models remains an important research question, especially for tasks that support critical decision-making and could affect minorities. The contribution of this work is a novel analysis covering architectures and embedding spaces for fine-grained understanding of bias over three approaches: generators, attribute modifier, and post-processing bias mitigators. This work shows that generators suffer from bias across all social groups with attribute preferences such as between 75%-85% for whiteness and 60%-80% for the female gender (for all trained CelebA models) and low probabilities of generating children and older men. Modifier and mitigators work as post-processor and change the generator performance. For instance, attribute channel perturbation strategies modify the embedding spaces. We quantify the influence of this change on group fairness by measuring the impact on image quality and group features. Specifically, we use the Fr\'echet Inception Distance (FID), the Face Matching Error and the Self-Similarity score. For Interfacegan, we analyze one and two attribute channel perturbations and examine the effect on the fairness distribution and the quality of the image. Finally, we analyzed the post-processing bias mitigators, which are the fastest and most computationally efficient way to mitigate bias. We find that these mitigation techniques show similar results on KL divergence and FID score, however, self-similarity scores show a different feature concentration on the new groups of the data distribution. The weaknesses and ongoing challenges described in this work must be considered in the pursuit of creating fair and unbiased face generation models.
翻译:GANs 和 扩散模型的最近进步使得高分辨率、超现实化图像得以生成。 但是,这些模型可能扭曲某些社会群体并显示偏差。 理解这些模型中的偏差仍然是一个重要的研究问题, 特别是对于支持关键决策并可能影响少数群体的任务而言。 这项工作的贡献是新颖的分析, 包括结构以及嵌入空间, 以便细化理解三种方法的偏差: 生成器、 属性修正器和后处理偏差减缓器。 这项工作表明, 产生者在所有社会群体中都受到偏差, 属性偏差, 例如白度为75- 85%, 女性性别偏差为60- 80 % 。 理解这些模型中的偏差仍然是一个重要的研究问题, 特别是对于产生子女和年长男子的低概率。 变差和减轻器作为后处理器的工作, 分化器调整嵌入空间。 我们量化这一变化对群体公平性的影响, 通过测量图像质量和群体特性的偏差。 具体地说, 我们使用Fr\ echet Incepion Condition Condition Legal (F, 所有Clegration rodeal cal rodeal) 解释女性性别分布的偏差, 和变差,, 和变变变差, 和变差的计算方法显示我们不断 和变错和变变变变变变变的 分析 的 和变法 的 的 。