Mixture of experts (MoE) is a popular class of models in statistics and machine learning that has sustained attention over the years, due to its flexibility and effectiveness. We consider the Gaussian-gated localized MoE (GLoME) regression model for modeling heterogeneous data. This model poses challenging questions with respect to the statistical estimation and model selection problems, including feature selection, both from the computational and theoretical points of view. We study the problem of estimating the number of components of the GLoME model, in a penalized maximum likelihood estimation framework. We provide a lower bound on the penalty that ensures a weak oracle inequality is satisfied by our estimator. To support our theoretical result, we perform numerical experiments on simulated and real data, which illustrate the performance of our finite-sample oracle inequality.
翻译:专家混合(MoE)是统计和机器学习方面最受欢迎的模型,多年来因其灵活性和有效性而一直受到关注。我们认为高山化本地化的MOE(GLOME)回归模型用于建模多种数据。这一模型对统计估计和模型选择问题提出了具有挑战性的问题,包括从计算和理论角度选择特征。我们研究了在受处罚的最大可能性估计框架内估算GLOME模型组成部分数量的问题。我们对于确保我们的估算者满足弱骨骼不平等的处罚提供了较低的约束。为了支持我们的理论结果,我们对模拟和真实数据进行了数字实验,这显示了我们有限的标本或标本不平等的表现。