We advocate for a practical Maximum Likelihood Estimation (MLE) approach towards designing loss functions for regression and forecasting, as an alternative to the typical approach of direct empirical risk minimization on a specific target metric. The MLE approach is better suited to capture inductive biases such as prior domain knowledge in datasets, and can output post-hoc estimators at inference time that can optimize different types of target metrics. We present theoretical results to demonstrate that our approach is competitive with any estimator for the target metric under some general conditions. In two example practical settings, Poisson and Pareto regression, we show that our competitive results can be used to prove that the MLE approach has better excess risk bounds than directly minimizing the target metric. We also demonstrate empirically that our method instantiated with a well-designed general purpose mixture likelihood family can obtain superior performance for a variety of tasks across time-series forecasting and regression datasets with different data distributions.
翻译:我们主张在设计回归和预测的损失函数时,采用实际的“最大可能性估计”法(MLE)来设计用于回归和预测的损失函数,作为在特定目标指标上直接将风险降到最低的典型经验性风险的替代方法。 MLE法更适合捕捉感性偏差,如在数据集中先前的域内知识,并可以在推论时间输出后热测算器,以优化不同类型的目标指标。我们提出理论结果,以证明我们的方法在某些一般条件下与任何目标指标的估测器具有竞争力。在Poisson和Pareto这两个实例的实用环境中,我们证明我们的竞争性结果可以用来证明MLE法比直接尽量减少目标指标的超风险范围要好。我们还从经验上表明,我们的方法与精心设计的通用混合可能性可实现的组合可以实现不同时间序列预测和不同数据分布的回归数据集的各种任务的优异性表现。