Mixture models are widely used in modeling heterogeneous data populations. A standard approach of mixture modeling is to assume that the mixture component takes a parametric kernel form, while the flexibility of the model can be obtained by using a large or possibly unbounded number of such parametric kernels. In many applications, making parametric assumptions on the latent subpopulation distributions may be unrealistic, which motivates the need for nonparametric modeling of the mixture components themselves. In this paper we study finite mixtures with nonparametric mixture components, using a Bayesian nonparametric modeling approach. In particular, it is assumed that the data population is generated according to a finite mixture of latent component distributions, where each component is endowed with a Bayesian nonparametric prior such as the Dirichlet process mixture. We present conditions under which the individual mixture component's distributions can be identified, and establish posterior contraction behavior for the data population's density, as well as densities of the latent mixture components. We develop an efficient MCMC algorithm for posterior inference and demonstrate via simulation studies and real-world data illustrations that it is possible to efficiently learn complex distributions for the latent subpopulations. In theory, the posterior contraction rate of the component densities is nearly polynomial, which is a significant improvement over the logarithm convergence rate of estimating mixing measures via deconvolution.
翻译:混合模型广泛应用于异质数据群体的建模。标准的混合建模方法通常假设混合成分具有参数核形式,而模型的灵活性可通过使用大量甚至无限数量的此类参数核实现。然而在许多应用中,对潜在子群体分布施加参数化假设可能不切实际,这促使我们需要对混合成分本身进行非参数建模。本文采用贝叶斯非参数建模方法,研究具有非参数混合成分的有限混合模型。具体而言,我们假设数据群体由有限个潜在成分分布混合生成,其中每个成分被赋予贝叶斯非参数先验(如狄利克雷过程混合)。我们提出了可识别各混合成分分布的条件,并建立了数据群体密度及潜在混合成分密度的后验收缩行为理论。我们开发了一种高效的马尔可夫链蒙特卡洛算法进行后验推断,通过模拟研究和真实数据示例证明:该方法能够有效学习潜在子群体的复杂分布。理论上,成分密度的后验收缩速率接近多项式阶,这相较于通过解卷积估计混合测度所获得的对数收敛速率有显著提升。