There is a consensus that focusing only on accuracy in searching for optimal machine learning models amplifies biases contained in the data, leading to unfair predictions and decision supports. Recently, multi-objective hyperparameter optimization has been proposed to search for machine learning models which offer equally Pareto-efficient trade-offs between accuracy and fairness. Although these approaches proved to be more versatile than fairness-aware machine learning algorithms -- which optimize accuracy constrained to some threshold on fairness -- they could drastically increase the energy consumption in the case of large datasets. In this paper we propose FanG-HPO, a Fair and Green Hyperparameter Optimization (HPO) approach based on both multi-objective and multiple information source Bayesian optimization. FanG-HPO uses subsets of the large dataset (aka information sources) to obtain cheap approximations of both accuracy and fairness, and multi-objective Bayesian Optimization to efficiently identify Pareto-efficient machine learning models. Experiments consider two benchmark (fairness) datasets and two machine learning algorithms (XGBoost and Multi-Layer Perceptron), and provide an assessment of FanG-HPO against both fairness-aware machine learning algorithms and hyperparameter optimization via a multi-objective single-source optimization algorithm in BoTorch, a state-of-the-art platform for Bayesian Optimization.
翻译:人们一致认为,在寻找最佳机器学习模型时只注重精确度,会扩大数据中包含的偏差,从而导致不公平的预测和决定支持。最近,提出了多目标超参数优化,以寻找在准确性和公平性之间提供同样Pareto效率权衡的机器学习模型。虽然这些方法证明比公平、了解的机器学习算法(这种算法将精度优化到一定的公平门槛)更为多样,但它们可以大幅提高大型数据集的能源消耗量。在本文中,我们建议采用基于多目标和多种信息来源Bayesian优化的FanG-HPO(公平与绿色超参数优化)公平与绿色超参数优化(HPO)方法。FanG-HPO利用大型数据集(aka信息来源)的子集来获取准确和公平性两者的廉价近似值,以及多目标Bayesian Oppimization,以有效识别Pareto-e-egal 机器学习模型。实验认为,两种基准(公平性)数据集和两种机器学习算法(XGBst和多-Ler- Percepron)的方法方法,并且对Fan-G-ral-Sqal-Sqal-Sqal-IAR-Sq-Sqal-Sqal-Sqal-Sqal-Sqal-Sqal-Sqal-Sq-Sq-Sal-I-S-SAL-SAR-A-SAR-A-S-SAR-A-S-SAR-SAR-S-SAR-S-S-SAR-A-S-S-S-S-S-S-S-S-S-S-SBA-S-SAR-S-S-S-S-S-S-S-S-S-A-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-Sir-A-SAR-S-S-S-A-A-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-