This work tackles the issue of fairness in the context of generative procedures, such as image super-resolution, which entail different definitions from the standard classification setting. Moreover, while traditional group fairness definitions are typically defined with respect to specified protected groups -- camouflaging the fact that these groupings are artificial and carry historical and political motivations -- we emphasize that there are no ground truth identities. For instance, should South and East Asians be viewed as a single group or separate groups? Should we consider one race as a whole or further split by gender? Choosing which groups are valid and who belongs in them is an impossible dilemma and being "fair" with respect to Asians may require being "unfair" with respect to South Asians. This motivates the introduction of definitions that allow algorithms to be \emph{oblivious} to the relevant groupings. We define several intuitive notions of group fairness and study their incompatibilities and trade-offs. We show that the natural extension of demographic parity is strongly dependent on the grouping, and \emph{impossible} to achieve obliviously. On the other hand, the conceptually new definition we introduce, Conditional Proportional Representation, can be achieved obliviously through Posterior Sampling. Our experiments validate our theoretical results and achieve fair image reconstruction using state-of-the-art generative models.
翻译:这项工作解决了在基因化程序背景下的公平问题,例如图像超分辨率,这需要与标准分类设置不同的定义。此外,传统群体公平定义通常被特定受保护群体所定义 -- -- 掩盖这些群体是人为的,带有历史和政治动机的事实 -- -- 我们强调,不存在任何事实真相特征。例如,南亚和东亚人是否应当被视为一个单一群体或单独的群体?我们是否认为一个种族是整个种族还是进一步被性别进一步分裂?选择哪些群体有效,谁属于这些群体,对于亚洲人来说,是一种不可能的两难境地,而且“公平”可能要求对南亚人来说是“不公平的”。 这促使采用允许这些群体进行算法的逻辑,使相关群体能够成为emph{obliviews。我们定义了几个关于群体公平性的直观概念,并研究其不相容性和权衡性。我们表明,人口均等的自然延伸在很大程度上取决于组别,谁属于这些群体是无法被接受的两难境地,对于亚洲人来说,可能是一种“不公平”的两难困境。这促使采用“不公平”对南亚人进行“不公平”的定义。这需要引入使算算算算法,而我们通过另一个结构的模型的模型的模型,我们实现了另一个结构上的模型的模型的重建。我们实现。我们所实现的模型。