In our experience of working with domain experts who are using today's AutoML systems, a common problem we encountered is what we call "unrealistic expectations" -- when users are facing a very challenging task with noisy data acquisition process, whilst being expected to achieve startlingly high accuracy with machine learning (ML). Consequently, many computationally expensive AutoML runs and labour-intensive ML development processes are predestined to fail from the beginning. In traditional software engineering, this problem is addressed via a feasibility study, an indispensable step before developing any software system. In this paper, we present ease.ml/snoopy with the goal of preforming an automatic feasibility study before building ML applications or collecting too many samples. A user provides inputs in the form of a dataset, which is representative for the task and data acquisition process, and a quality target (e.g., expected accuracy > 0.8). The system returns its deduction on whether this target is achievable using ML given the input data. We approach this problem by estimating the irreducible error of the underlying task, also known as Bayes error. The technical key contribution of this work is the design of a practical Bayes error estimator. We carefully evaluate the benefits and limitations of running ease.ml/snoopy prior to training ML models on too noisy datasets for reaching the desired target accuracy. By including the automatic feasibility study into the iterative label cleaning process, users are able to save substantial labeling time and monetary efforts.
翻译:在与使用今天的自动ML系统的域专家一起工作的经验中,我们遇到的一个共同问题是我们所谓的“不现实期望”——当用户在繁忙的数据获取过程中面临非常艰巨的任务时,当用户面临一个非常艰巨的任务时,在机器学习(ML)时,预期会达到惊人的高精度。因此,许多计算成本昂贵的自动ML运行和劳力密集型ML开发过程注定从一开始就失败。在传统的软件工程中,这个问题是通过可行性研究来解决的,这是开发任何软件系统之前一个不可或缺的步骤。在本文中,我们提出“不现实的期望”,目的是在建立ML应用程序或收集过多样本之前预先进行自动可行性研究。一个用户以数据集的形式提供投入,该数据集对任务和数据获取过程和质量目标(例如,预期准确度 > 0.8)具有代表性。在传统的软件工程中,系统通过利用ML数据来推算这个目标是否可以实现。我们通过估计基本任务的不可避免的反复反复性错误来解决这个问题,也称为Bayes错误。我们的技术关键工作的关键贡献是设计一个实际错误的准确性模型。