There has been increasing awareness of ethical issues in machine learning, and fairness has become an important research topic. Most fairness efforts in computer vision have been focused on human sensing applications and preventing discrimination by people's physical attributes such as race, skin color or age by increasing visual representation for particular demographic groups. We argue that ML fairness efforts should extend to object recognition as well. Buildings, artwork, food and clothing are examples of the objects that define human culture. Representing these objects fairly in machine learning datasets will lead to models that are less biased towards a particular culture and more inclusive of different traditions and values. There exist many research datasets for object recognition, but they have not carefully considered which classes should be included, or how much training data should be collected per class. To address this, we propose a simple and general approach, based on crowdsourcing the demographic composition of the contributors: we define fair relevance scores, estimate them, and assign them to each class. We showcase its application to the landmark recognition domain, presenting a detailed analysis and the final fairer landmark rankings. We present analysis which leads to a much fairer coverage of the world compared to existing datasets. The evaluation dataset was used for the 2021 Google Landmark Challenges, which was the first of a kind with an emphasis on fairness in generic object recognition.
翻译:对机器学习中的伦理问题的认识不断提高,公平已成为一个重要的研究课题。计算机视野中的大多数公平努力都侧重于人类遥感应用,并通过提高特定人口群体的视觉代表性,防止因种族、肤色或年龄等人的身体属性而受种族、肤色或年龄歧视。我们认为,建筑、艺术品、食品和服装是界定人类文化的物体的例子。在机器学习数据集中公平代表这些物体将导致模型,这些模型较少偏向特定文化,更包容不同的传统和价值观。有许多研究数据集用于物体识别,但没有仔细考虑应列入哪些类,或每类应收集多少培训数据。为了解决这个问题,我们提议一个简单和一般的方法,根据人群划分贡献者的人口构成:我们界定公平的相关性分数,估计它们,并将其分配给每一类。我们展示其在标志性识别域的应用情况,提出详细分析,最后更公平的标定等级。我们的分析导致比现有数据集更公平的覆盖世界范围,但没有仔细考虑应列入哪些类,或者每类应收集多少培训数据。为了解决这一问题,我们提出了一种简单和笼统的方法,即评估2021年谷域中所使用的数据定义了一种标准。