From a model-building perspective, in this paper we propose a paradigm shift for fitting over-parameterized models. Philosophically, the mindset is to fit models to future observations rather than to the observed sample. Technically, choosing an imputation model for generating future observations, we fit over-parameterized models to future observations via optimizing an approximation to the desired expected loss-function based on its sample counterpart and an adaptive simplicity-preference function. This technique is discussed in detail to both creating bootstrap imputation and final estimation with bootstrap imputation. The method is illustrated with the many-normal-means problem, $n < p$ linear regression, and deep convolutional neural networks for image classification of MNIST digits. The numerical results demonstrate superior performance across these three different types of applications. For example, for the many-normal-means problem, our method uniformly dominates James-Stein and Efron's $g-$modeling, and for the MNIST image classification, it performs better than all existing methods and reaches arguably the best possible result. While this paper is largely expository because of the ambitious task of taking a look at over-parameterized models from the new perspective, fundamental theoretical properties are also investigated. We conclude the paper with a few remarks.
翻译:从建模的角度来看,本文中我们建议对设计过度参数模型进行范式转变。从哲学角度讲,这种思维模式是将模型与未来观测而不是观察到的样本相适应。从技术上讲,选择一种估算模型来生成未来观测,我们通过优化对基于抽样对应方的预期损失功能的近似和适应性简单偏重功能,将过度量化模型与未来观测相匹配。我们详细讨论了这一技术,既要创建靴子陷阱估算和用靴子陷阱套图进行最后估计。该方法与许多正常手段问题($ < p$ 线性回归)和用于MMIST数字图像分类的深革命性神经网络进行了演示。数字结果显示了这三种不同应用类型的优异性表现。例如,对于许多正常手段问题,我们的方法一致地支配着James-Stein和Efron的美元模型模型,对于MNISTS图像分类来说,它比所有现有方法都好,并且达到了最有说服力的结果。虽然本文基本上是从理论角度来解释的,因为我们从一个雄心勃勃的论文的模型的角度来看,我们也是从一个结论。