Modern statistics provides an ever-expanding toolkit for estimating unknown parameters. Consequently, applied statisticians frequently face a difficult decision: retain a parameter estimate from a familiar method or replace it with an estimate from a newer or complex one. While it is traditional to compare estimators using risk, such comparisons are rarely conclusive in realistic settings. In response, we propose the "c-value" as a measure of confidence that a new estimate achieves smaller loss than an old estimate on a given dataset. We show that it is unlikely that a computed c-value is large and that the new estimate has larger loss than the old. Therefore, just as a small p-value provides evidence to reject a null hypothesis, a large c-value provides evidence to use a new estimate in place of the old. For a wide class of problems and estimators, we show how to compute a c-value by first constructing a data-dependent high-probability lower bound on the difference in loss. The c-value is frequentist in nature, but we show that it can provide a validation of Bayesian estimates in real data applications involving hierarchical models and Gaussian processes.
翻译:现代统计为估计未知参数提供了一个不断扩大的工具包。 因此,应用统计人员经常面临一个困难的决定:保留一种熟悉方法的参数估计,或用一种较新或较复杂方法的估计数来取代这种估计。虽然传统上比较使用风险的估算者,但这种比较在现实环境中很少是结论性的。作为回应,我们建议“c-value”作为一种信任度的衡量标准,即新估计的损失小于对某一数据集的旧估计。我们表明,计算出的C-value不可能很大,新的估计损失大于旧的。因此,一个小的p-value提供了拒绝无效假设的证据,而一个大的c-value则提供了在旧的假设中使用新估计的证据。对于广泛的问题和估计者来说,我们展示了如何通过首先构建一种依赖数据的高概率对损失差异的较低约束来计算C-value。 c-value是经常发生的,但我们表明,它可以在涉及等级模型和Gassian进程的真实数据应用中验证Bayesian的估计。