Virtually all machine learning tasks are characterized using some form of loss function, and "good performance" is typically stated in terms of a sufficiently small average loss, taken over the random draw of test data. While optimizing for performance on average is intuitive, convenient to analyze in theory, and easy to implement in practice, such a choice brings about trade-offs. In this work, we survey and introduce a wide variety of non-traditional criteria used to design and evaluate machine learning algorithms, place the classical paradigm within the proper historical context, and propose a view of learning problems which emphasizes the question of "what makes for a desirable loss distribution?" in place of tacit use of the expected loss.
翻译:几乎所有的机器学习任务都使用某种形式的损失功能来定性,而“良好业绩”通常以足够小的平均损失来表述,取而代之的是随机抽取测试数据。 平均而言,优化性能是直观的,便于在理论上分析,在实践中易于执行,但这种选择会带来权衡。 在这项工作中,我们调查并引入了各种非传统标准,用于设计和评估机器学习算法,将经典模式置于适当的历史背景中,并提出学习问题观点,强调“什么导致适当的损失分配”的问题,以取代对预期损失的默认使用。