We address an inherent difficulty in welfare-theoretic fair machine learning by proposing an equivalently axiomatically-justified alternative and studying the resulting computational and statistical learning questions. Welfare metrics quantify overall wellbeing across a population of one or more groups, and welfare-based objectives and constraints have recently been proposed to incentivize fair machine learning methods to produce satisfactory solutions that consider the diverse needs of multiple groups. Unfortunately, many machine-learning problems are more naturally cast as loss minimization tasks, rather than utility maximization, which complicates direct application of welfare-centric methods to fair machine learning. In this work, we define a complementary measure, termed malfare, measuring overall societal harm (rather than wellbeing), with axiomatic justification via the standard axioms of cardinal welfare. We then cast fair machine learning as malfare minimization over the risk values (expected losses) of each group. Surprisingly, the axioms of cardinal welfare (malfare) dictate that this is not equivalent to simply defining utility as negative loss. Building upon these concepts, we define fair-PAC (FPAC) learning, where an FPAC learner is an algorithm that learns an $\varepsilon$-$\delta$ malfare-optimal model with bounded sample complexity, for any data distribution, and for any (axiomatically justified) malfare concept. Finally, we show broad conditions under which, with appropriate modifications, standard PAC-learners may be converted to FPAC learners. This places FPAC learning on firm theoretical ground, as it yields statistical and computational efficiency guarantees for many well-studied machine-learning models, and is also practically relevant, as it democratizes fair ML by providing concrete training algorithms and rigorous generalization guarantees for these models
翻译:我们处理福利理论公平机器学习的内在困难,方法是提出一个相当有理的、有理有据的替代方法,并研究由此得出的计算和统计学习问题。福利衡量标准量化了一个或一个以上群体的总体福祉,最近提出了基于福利的目标和制约因素,以激励公平的机器学习方法,从而产生考虑到多个群体不同需要的满意解决方案。不幸的是,许多机器学习问题自然地被描绘成最大限度地减少损失的任务,而不是尽量扩大效用,这使得直接应用以福利为中心的方法进行公平的机器学习更为复杂。在这项工作中,我们定义了一个补充性措施,称为恶意,衡量总体社会伤害(而不是福利 ) 。 福利衡量标准指标衡量一个或一个以上群体的总体福祉,而基于福利基准xxxxxxxxxxx,我们提出公平的机器学习错误,而基于成本xxxxxxxxxxxxx的数学模型, 也显示成本xxxxxxxxx的数学分布。 令人惊讶的是,基本福利(错误) 基本福利(错误)意味着它不是简单地将实用的实用的、标准学习标准学习模式定义为成本的数学。