Twitter uses machine learning to crop images, where crops are centered around the part predicted to be the most salient. In fall 2020, Twitter users raised concerns that the automated image cropping system on Twitter favored light-skinned over dark-skinned individuals, as well as concerns that the system favored cropping woman's bodies instead of their heads. In order to address these concerns, we conduct an extensive analysis using formalized group fairness metrics. We find systematic disparities in cropping and identify contributing factors, including the fact that the cropping based on the single most salient point can amplify the disparities because of an effect we term argmax bias. However, we demonstrate that formalized fairness metrics and quantitative analysis on their own are insufficient for capturing the risk of representational harm in automatic cropping. We suggest the removal of saliency-based cropping in favor of a solution that better preserves user agency. For developing a new solution that sufficiently address concerns related to representational harm, our critique motivates a combination of quantitative and qualitative methods that include human-centered design.
翻译:在2020年秋季,推特用户提出了如下关切:推特上的自动图像裁剪系统偏向于黑皮肤人的浅皮肤,以及该系统偏向于裁剪妇女的身体而不是她们的头部。为了解决这些问题,我们利用正式的团体公平度量度进行了广泛的分析。我们发现在作物种植方面存在系统性差异,并找出了造成差异的因素,包括基于单一最突出点的作物种植能够扩大差异,因为我们称之为“强力偏差”的影响。然而,我们表明,正式的公平度量度和定量分析本身不足以捕捉到在自动裁剪中造成代表性伤害的风险。我们建议删除突出的作物,以有利于更好地保护用户机构的解决办法。为了制定能够充分解决与代表性伤害有关的关切的新解决方案,我们的批评激励了包括以人为中心的设计在内的定量和定性方法的组合。