Thanks to the great progress of machine learning in the last years, several Artificial Intelligence (AI) techniques have been increasingly moving from the controlled research laboratory settings to our everyday life. AI is clearly supportive in many decision-making scenarios, but when it comes to sensitive areas such as health care, hiring policies, education, banking or justice, with major impact on individuals and society, it becomes crucial to establish guidelines on how to design, develop, deploy and monitor this technology. Indeed the decision rules elaborated by machine learning models are data-driven and there are multiple ways in which discriminatory biases can seep into data. Algorithms trained on those data incur the risk of amplifying prejudices and societal stereotypes by over associating protected attributes such as gender, ethnicity or disabilities with the prediction task. Starting from the extensive experience of the National Metrology Institute on measurement standards and certification roadmaps, and of Politecnico di Torino on machine learning as well as methods for domain bias evaluation and mastering, we propose a first joint effort to define the operational steps needed for AI fairness certification. Specifically we will overview the criteria that should be met by an AI system before coming into official service and the conformity assessment procedures useful to monitor its functioning for fair decisions.
翻译:过去几年来,由于机器学习取得了巨大进步,人工智能(AI)技术从受控制的研究实验室环境逐渐转向我们的日常生活,人工智能(AI)技术在许多决策情景中显然支持许多决策情景,但当涉及保健、雇用政策、教育、银行或司法等敏感领域,对个人和社会产生重大影响时,必须制定如何设计、开发、部署和监测这种技术的准则,机器学习模型所制定的决策规则确实以数据为驱动因素,歧视性偏见可以以多种方式渗入数据。受过关于这些数据的培训的人工智能有可能将性别、族裔或残疾等受保护属性与预测任务联系起来,从而扩大偏见和社会陈规定型观念。从国家计量研究所关于衡量标准和认证路线图的广泛经验,以及Politecnico di Torino关于机器学习以及域偏差评价和掌握方法的广泛经验,我们提议首次共同努力确定AI公平认证所需的操作步骤。我们将在进入正式服务和运行符合性评估程序之前,先通过一个AI系统来概述应达到的标准。