This paper demonstrates the systematic use of combinatorial coverage for selecting and characterizing test and training sets for machine learning models. The presented work adapts combinatorial interaction testing, which has been successfully leveraged in identifying faults in software testing, to characterize data used in machine learning. The MNIST hand-written digits data is used to demonstrate that combinatorial coverage can be used to select test sets that stress machine learning model performance, to select training sets that lead to robust model performance, and to select data for fine-tuning models to new domains. Thus, the results posit combinatorial coverage as a holistic approach to training and testing for machine learning. In contrast to prior work which has focused on the use of coverage in regard to the internal of neural networks, this paper considers coverage over simple features derived from inputs and outputs. Thus, this paper addresses the case where the supplier of test and training sets for machine learning models does not have intellectual property rights to the models themselves. Finally, the paper addresses prior criticism of combinatorial coverage and provides a rebuttal which advocates the use of coverage metrics in machine learning applications.
翻译:本文展示了为机械学习模式选择和定性测试和培训成套材料的分类覆盖的系统使用; 介绍的工作调整了组合互动测试,在确定软件测试的缺陷方面成功地利用了这种测试来描述机器学习中所使用的数据; MNIST手写的数字数据用于证明可使用组合覆盖来选择压力机器学习模型性能的测试组,选择能够产生稳健模型性能的培训组,并为微调模型选择数据到新领域。 因此,结果将组合覆盖作为培训和测试机器学习的整体方法。 与以前侧重于使用神经网络内部覆盖的工作相比,本文考虑了对输入和产出产生的简单特征的覆盖。 因此,本文述及了机器学习模型的测试和培训组供应商本身不具有知识产权的情况。 最后,本文述及了对组合覆盖范围的先前批评,并提供了一种反驳,主张在机器学习应用中使用覆盖指标。