Ad platforms such as Facebook, Google and LinkedIn promise value for advertisers through their targeted advertising. However, multiple studies have shown that ad delivery on such platforms can be skewed by gender or race due to hidden algorithmic optimization by the platforms, even when not requested by the advertisers. Building on prior work measuring skew in ad delivery, we develop a new methodology for black-box auditing of algorithms for discrimination in the delivery of job advertisements. Our first contribution is to identify the distinction between skew in ad delivery due to protected categories such as gender or race, from skew due to differences in qualification among people in the targeted audience. This distinction is important in U.S. law, where ads may be targeted based on qualifications, but not on protected categories. Second, we develop an auditing methodology that distinguishes between skew explainable by differences in qualifications from other factors, such as the ad platform's optimization for engagement or training its algorithms on biased data. Our method controls for job qualification by comparing ad delivery of two concurrent ads for similar jobs, but for a pair of companies with different de facto gender distributions of employees. We describe the careful statistical tests that establish evidence of non-qualification skew in the results. Third, we apply our proposed methodology to two prominent targeted advertising platforms for job ads: Facebook and LinkedIn. We confirm skew by gender in ad delivery on Facebook, and show that it cannot be justified by differences in qualifications. We fail to find skew in ad delivery on LinkedIn. Finally, we suggest improvements to ad platform practices that could make external auditing of their algorithms in the public interest more feasible and accurate.
翻译:Facebook、 Google 和 LinkedIn 等平台。 然而,多项研究表明,由于平台隐藏的算法优化,即使广告商没有提出要求,这些平台上的广告可能会由于性别或种族差异而被性别或种族扭曲。 然而,许多研究表明,由于平台隐藏的逻辑优化,这些平台上的广告可能会被性别或种族扭曲。基于先前的工作,我们开发了一种新的方法,用于对在提供招聘广告时歧视的算法进行黑箱审计。我们的第一个贡献是确定由于受保护类别(如性别或种族)而导致的在交付中出现偏差的区别。由于目标受众在资格方面存在差异,这种区别在此类平台上可能因性别或种族而出现偏差。在美国法律中,这种区别很重要,因为平台的广告可能是基于资格的隐蔽的逻辑优化,而不是针对受保护的类别。 其次,我们开发了一种审计方法,根据资格与其他因素的不同,比如广告平台在参与或培训有偏差的数据中优化。 我们通过将两种同时交付的广告平台与在交付中的不准确的广告平台进行对比,我们无法对等公司进行合理的调整。