In this article I propose an approach for defining replicability for prediction rules. Motivated by a recent NAS report, I start from the perspective that replicability is obtaining consistent results across studies suitable to address the same prediction question, each of which has obtained its own data. I then discuss concept and issues in defining key elements of this statement. I focus specifically on the meaning of "consistent results" in typical utilization contexts, and propose a multi-agent framework for defining replicability, in which agents are neither partners nor adversaries. I recover some of the prevalent practical approaches as special cases. I hope to provide guidance for a more systematic assessment of replicability in machine learning.
翻译:翻译摘要:
在本文中,我提出了一种方法来定义预测规则的可复制性。受近期美国国家科学院报告的启发,我从每个研究都有自己的数据,适合解决相同预测问题的角度出发。我然后讨论定义该语句关键要素的概念和问题。我特别关注在典型利用环境下“一致结果”的含义,并提出了一个多智能体框架来定义可复制性,其中,智能体既不是合作伙伴也不是对手。我恢复了一些主要的实际方法作为特例。我希望为机器学习中更系统的可复制性评估提供指导。