Generative Adversarial Networks (GANs) are capable of synthesizing high-quality facial images. Despite their success, GANs do not provide any information about the relationship between the input vectors and the generated images. Currently, facial GANs are trained on imbalanced datasets, which generate less diverse images. For example, more than 77% of 100K images that we randomly synthesized using the StyleGAN3 are classified as Happy, and only around 3% are Angry. The problem even becomes worse when a mixture of facial attributes is desired: less than 1% of the generated samples are Angry Woman, and only around 2% are Happy Black. To address these problems, this paper proposes a framework, called GANalyzer, for the analysis, and manipulation of the latent space of well-trained GANs. GANalyzer consists of a set of transformation functions designed to manipulate latent vectors for a specific facial attribute such as facial Expression, Age, Gender, and Race. We analyze facial attribute entanglement in the latent space of GANs and apply the proposed transformation for editing the disentangled facial attributes. Our experimental results demonstrate the strength of GANalyzer in editing facial attributes and generating any desired faces. We also create and release a balanced photo-realistic human face dataset. Our code is publicly available on GitHub.
翻译:例如,我们使用StyleGAN3随机合成的100K图像中有77%以上被归类为“快乐”,只有大约3%被归为“愤怒”。在需要组合面部特征时,问题甚至更加严重:不到1%的样本是愤怒女性,只有2%左右是快乐黑人。为了解决这些问题,本文提出了一个框架,称为GaNalyzer,用于分析并操纵受过良好训练的GANs的潜在空间。Galyzer由一套转换功能组成,旨在操纵潜向矢量以取得面部表现、年龄、性别和种族等特定面部属性。我们分析了GANs潜在空间的面部特征,并应用了拟议转换来编辑我们所期望的面部和面部特征。我们制作的面部图像和面部特征的实验性GAlyzer也展示了我们所期望的面部特征。