Recent scholarship has argued that firms building data-driven decision systems in high-stakes domains like employment, credit, and housing should search for "less discriminatory algorithms" (LDAs) (Black et al., 2024). That is, for a given decision problem, firms considering deploying a model should make a good-faith effort to find equally performant models with lower disparate impact across social groups. Evidence from the literature on model multiplicity shows that randomness in training pipelines can lead to multiple models with the same performance, but meaningful variations in disparate impact. This suggests that developers can find LDAs simply by randomly retraining models. Firms cannot continue retraining forever, though, which raises the question: What constitutes a good-faith effort? In this paper, we formalize LDA search via model multiplicity as an optimal stopping problem, where a model developer with limited information wants to produce strong evidence that they have sufficiently explored the space of models. Our primary contribution is an adaptive stopping algorithm that yields a high-probability upper bound on the gains achievable from a continued search, allowing the developer to certify (e.g., to a court) that their search was sufficient. We provide a framework under which developers can impose stronger assumptions about the distribution of models, yielding correspondingly stronger bounds. We validate the method on real-world credit, employment and housing datasets.
翻译:近期研究指出,在就业、信贷和住房等高风险领域构建数据驱动决策系统的企业,应当寻求"更少歧视性的算法"(Black等人,2024)。这意味着针对特定决策问题,企业在考虑部署模型时,应本着诚信原则努力寻找性能相当但对社会群体间差异影响更低的模型。模型多重性相关文献表明,训练流程中的随机性可能导致多个性能相同但差异影响程度存在显著差异的模型。这表明开发者仅通过随机重新训练模型即可找到更少歧视性算法。然而企业无法无限期持续重新训练,这引出一个核心问题:何种程度可被视为诚信努力?本文通过模型多重性将更少歧视性算法搜索形式化为最优停止问题,其中信息有限的模型开发者需要提供充分证据,证明其已对模型空间进行了足够探索。我们的主要贡献是提出一种自适应停止算法,该算法能以高概率给出持续搜索可获收益的上界,使开发者能够(例如向法庭)证明其搜索充分性。我们构建了一个框架,开发者可在该框架下对模型分布施加更强假设,从而获得相应更强的边界保证。我们在真实世界的信贷、就业和住房数据集上验证了该方法的有效性。