Nowadays model uncertainty has become one of the most important problems in both academia and industry. In this paper, we mainly consider the scenario in which we have a common model set used for model averaging instead of selecting a single final model via a model selection procedure to account for this model's uncertainty to improve the reliability and accuracy of inferences. Here one main challenge is to learn the prior over the model set. To tackle this problem, we propose two data-based algorithms to get proper priors for model averaging. One is for meta-learner, the analysts should use historical similar tasks to extract the information about the prior. The other one is for base-learner, a subsampling method is used to deal with the data step by step. Theoretically, an upper bound of risk for our algorithm is presented to guarantee the performance of the worst situation. In practice, both methods perform well in simulations and real data studies, especially with poor-quality data.
翻译:目前,模型不确定性已成为学术界和产业界最重要的问题之一。在本文中,我们主要考虑一种假设,即我们有一个共同的模型集,用于平均模型,而不是通过一个模型选择程序选择一个单一的最后模型,以计算模型的不确定性,从而提高推论的可靠性和准确性。这里的一个主要挑战是先学习模型集的可靠性和准确性。为了解决这一问题,我们建议采用两种基于数据的算法,以获得模型平均的正确前科。一个是元利阿尔,分析员应使用类似的历史任务来提取关于前一数据的信息。另一个是基利阿尔,用一种次级抽样方法逐步处理数据。理论上,我们算法的风险上限用来保证最坏的情况。在实践中,两种方法在模拟和真实数据研究中都运作良好,特别是质量差的数据。