Rapid advancements in artificial intelligence (AI) technology have brought about a plethora of new challenges in terms of governance and regulation. AI systems are being integrated into various industries and sectors, creating a demand from decision-makers to possess a comprehensive and nuanced understanding of the capabilities and limitations of these systems. One critical aspect of this demand is the ability to explain the results of machine learning models, which is crucial to promoting transparency and trust in AI systems, as well as fundamental in helping machine learning models to be trained ethically. In this paper, we present novel quantitative metrics frameworks for interpreting the predictions of classifier and regressor models. The proposed metrics are model agnostic and are defined in order to be able to quantify: i. the interpretability factors based on global and local feature importance distributions; ii. the variability of feature impact on the model output; and iii. the complexity of feature interactions within model decisions. We employ publicly available datasets to apply our proposed metrics to various machine learning models focused on predicting customers' credit risk (classification task) and real estate price valuation (regression task). The results expose how these metrics can provide a more comprehensive understanding of model predictions and facilitate better communication between decision-makers and stakeholders, thereby increasing the overall transparency and accountability of AI systems.
翻译:人工智能(AI)技术的迅速发展带来了许多新的治理和监管挑战。人工智能系统正在被整合到各种行业和部门,使决策者要求对这些系统的能力和局限性有一个全面和细致的了解。这一要求的一个重要方面是能够解释机器学习模型的结果,这对于促进AI系统的透明度和信任至关重要,对于帮助机器学习模型接受道德培训至关重要。本文提出了新的定量指标框架,用于解释分类和累进模型预测。拟议指标是模型,定义是为了能够量化:即基于全球和地方特征重要性分布的可解释性因素;对模型产出的特征影响变化;以及模型决定中特征互动的复杂性。我们利用公开提供的数据集,将我们提出的指标应用于侧重于预测客户信用风险(分类任务)和房地产价格估价(倒退任务)的各种机器学习模型模型。拟议指标是模型的模型,其定义是为了能够量化:即基于全球和地方特征重要性分布的可解释性因素;二。这些特征对模型产出的影响变化;三. 模型决定中特征互动的复杂性。我们利用公开的数据集,将我们提出的指标应用于侧重于预测客户信用风险(分类任务)和房地产价格估价(倒退任务)。结果揭示结果揭示了提高透明度,从而有助于对AI决定系统进行更全面的预测。