Learning fair representation is crucial for achieving fairness or debiasing sensitive information. Most existing works rely on adversarial representation learning to inject some invariance into representation. However, adversarial learning methods are known to suffer from relatively unstable training, and this might harm the balance between fairness and predictiveness of representation. We propose a new approach, learning FAir Representation via distributional CONtrastive Variational AutoEncoder (FarconVAE), which induces the latent space to be disentangled into sensitive and nonsensitive parts. We first construct the pair of observations with different sensitive attributes but with the same labels. Then, FarconVAE enforces each non-sensitive latent to be closer, while sensitive latents to be far from each other and also far from the non-sensitive latent by contrasting their distributions. We provide a new type of contrastive loss motivated by Gaussian and Student-t kernels for distributional contrastive learning with theoretical analysis. Besides, we adopt a new swap-reconstruction loss to boost the disentanglement further. FarconVAE shows superior performance on fairness, pretrained model debiasing, and domain generalization tasks from various modalities, including tabular, image, and text.
翻译:大部分现有工作都依靠对抗性代表学习来将某些偏差引入代表制。然而,据知对抗性学习方法会受到相对不稳定的培训,这可能会损害代表制公平性和预测性之间的平衡。我们提出一种新的方法,即通过分布式、横向变化式自动编码(FarconVAE)学习Fair代表制(Fair Prespresent),让潜在空间与敏感和不敏感部分脱钩。我们首先用不同的敏感属性来构建观测组合,但标签相同。然后,FarconVAE实施每一种不敏感的潜在关系,使之更加接近,而敏感的潜在关系则通过对比其分布而远离对方,远离不敏感的潜在关系。我们提供了一种新型的对比性损失,由Gaussian和学生心心室(Farconvational Vative Auteconomical-T ) 和理论分析为分布式对比性学习提供动力。此外,我们采用新的互换式重组损失来进一步加深分歧。FarconvaE显示公平性、先变式、格式和域图案的优业绩,包括一般格式和域图案。