Increased adoption of artificial intelligence (AI) systems into scientific workflows will result in an increasing technical debt as the distance between the data scientists and engineers who develop AI system components and scientists, researchers and other users grows. This could quickly become problematic, particularly where guidance or regulations change and once-acceptable best practice becomes outdated, or where data sources are later discredited as biased or inaccurate. This paper presents a novel method for deriving a quantifiable metric capable of ranking the overall transparency of the process pipelines used to generate AI systems, such that users, auditors and other stakeholders can gain confidence that they will be able to validate and trust the data sources and contributors in the AI systems that they rely on. The methodology for calculating the metric, and the type of criteria that could be used to make judgements on the visibility of contributions to systems are evaluated through models published at ModelHub and PyTorch Hub, popular archives for sharing science resources, and is found to be helpful in driving consideration of the contributions made to generating AI systems and approaches towards effective documentation and improving transparency in machine learning assets shared within scientific communities.
翻译:随着开发AI系统组成部分的数据科学家和工程师与科学家、研究人员和其他用户之间的距离日益扩大,在科学工作流程中更多地采用人工智能系统将造成越来越多的技术债务,这可能会很快成为问题,特别是在指南或条例的改变和一旦可接受的最佳做法过时,或数据来源后来因偏差或不准确而名誉受损的情况下,本文件提出了一种新的方法,用以得出量化指标,能够对用于生成AI系统的进程管道的总体透明度进行排名,使用户、审计员和其他利益攸关方能够相信他们能够验证和信任他们所依赖的AI系统的数据来源和提供者。计算衡量标准的方法和可用于判断对系统所作贡献的能见度的标准类型,通过模型Hub和PyTorrch Hub公布的模式进行评估,这些模型是分享科学资源的流行档案,有助于推动考虑为创建AI系统所做的贡献,有助于在科学界共享的机器学习资产方面实现有效的文件记录和提高透明度。