Boosting methods are widely used in statistical learning to deal with high-dimensional data due to their variable selection feature. However, those methods lack straightforward ways to construct estimators for the precision of the parameters such as variance or confidence interval, which can be achieved by conventional statistical methods like Bayesian inference. In this paper, we propose a new inference method "BayesBoost" that combines boosting and Bayesian for linear mixed models to make the uncertainty estimation for the random effects possible on the one hand. On the other hand, the new method overcomes the shortcomings of Bayesian inference in giving precise and unambiguous guidelines for the selection of covariates by benefiting from boosting techniques. The implementation of Bayesian inference leads to the randomness of model selection criteria like the conditional AIC (cAIC), so we also propose a cAIC-based model selection criteria that focus on the stabilized regions instead of the global minimum. The effectiveness of the new approach can be observed via simulation and in a data example from the field of neurophysiology focussing on the mechanisms in the brain while listening to unpleasant sounds.
翻译:在统计学学习中广泛使用推导方法来处理高维数据,因为其选择特点各异。然而,这些方法缺乏直接的方法来构建参数精确度的测算器,例如差异或信任间隔,可以通过贝叶斯推论等常规统计方法实现。在本文中,我们提议一种新的推论方法“BayesBoost”,将推力和巴耶斯混合模型结合起来,以便对随机效应进行不确定性估计。另一方面,新方法克服了巴伊西亚推论的缺点,即通过提振技术为选择共变体提供准确和明确的指导方针。采用巴伊西亚推论,导致模型选择标准的随机性,如有条件的AIC(CAIC),因此我们还建议以CAIC为基础的模型选择标准,以稳定区域而不是全球最低值为重点。通过模拟和从神经物理学领域观察新方法的有效性,侧重于大脑中的机制,同时倾听不愉快的声音。