Following the recent surge in adoption of machine learning (ML), the negative impact that improper use of ML can have on users and society is now also widely recognised. To address this issue, policy makers and other stakeholders, such as the European Commission or NIST, have proposed high-level guidelines aiming to promote trustworthy ML (i.e., lawful, ethical and robust). However, these guidelines do not specify actions to be taken by those involved in building ML systems. In this paper, we argue that guidelines related to the development of trustworthy ML can be translated to operational practices, and should become part of the ML development life cycle. Towards this goal, we ran a multi-vocal literature review, and mined operational practices from white and grey literature. Moreover, we launched a global survey to measure practice adoption and the effects of these practices. In total, we identified 14 new practices, and used them to complement an existing catalogue of ML engineering practices. Initial analysis of the survey results reveals that so far, practice adoption for trustworthy ML is relatively low. In particular, practices related to assuring security of ML components have very low adoption. Other practices enjoy slightly larger adoption, such as providing explanations to users. Our extended practice catalogue can be used by ML development teams to bridge the gap between high-level guidelines and actual development of trustworthy ML systems; it is open for review and contribution
翻译:在最近采用机器学习(ML)的激增之后,人们现在也广泛认识到不当使用ML可能对用户和社会产生的消极影响,为了解决这个问题,决策者和其他利益相关者,例如欧洲委员会或NIST,提出了旨在促进可靠的ML(即合法、合乎道德和健全的)的高级别准则,然而,这些准则并未具体说明参与建立ML系统的人应采取的行动。在本文件中,我们争辩说,与开发可信赖的ML有关的准则可以转化为业务做法,并应成为ML发展生命周期的一部分。为了实现这一目标,我们进行了多voal文献审查,并从白灰文献中挖掘了业务做法。此外,我们发起了一项全球调查,以衡量采用这些做法的做法及其影响。我们总共确定了14项新做法,并用这些新做法补充了ML工程做法的现有目录。对调查结果的初步分析表明,迄今为止,采用可信赖ML做法的做法相对较低,特别是,与确保ML组成部分安全有关的做法的采用率很低。其他做法享有稍大得多的多语言审查,而且从白和灰文献中提取了业务做法,我们为用户提供可靠的ML级发展指南,这是我们为用户提供更可靠的参考性发展系统所使用的标准。