Bayesian hierarchical mixture clustering (BHMC) is an interesting model that improves on the traditional Bayesian hierarchical clustering approaches. Regarding the parent-to-node diffusion in the generative process, BHMC replaces the conventional Gaussian-to-Gaussian (G2G) kernels with a Hierarchical Dirichlet Process Mixture Model (HDPMM). However, the drawback of the BHMC lies in that it might obtain comparatively high nodal variance in the higher levels (i.e., those closer to the root node). This can be interpreted as that the separation between the nodes, in particular those in the higher levels, might be weak. Attempting to overcome this drawback, we consider a recent inferential framework named posterior regularization, which facilitates a simple manner to impose extra constraints on a Bayesian model to address some weakness of the original model. Hence, to enhance the separation of clusters, we apply posterior regularization to impose max-margin constraints on the nodes at every level of the hierarchy. In this paper, we illustrate how the framework integrates with the BHMC and achieves the desired improvements over the original model.
翻译:BHMC是一个有趣的模式,它改进了传统的Bayesian等级分组方法。关于基因过程的母对节的传播,BHMC用一个等级分分解过程混合模型(HDPMM)取代传统的Gaussian-G2G(G2G)内核。然而,BHMC的缺点在于它可能会在较高层次(即更接近根节点的)获得相对较高的节点差异。这可以解释为节点之间的分离,特别是较高层次的节点之间的分离可能很弱。为了克服这一缺陷,我们考虑最近的推论框架,即称为后台规范化,这便于对Bayesian模式施加额外的限制,以解决原始模式的某些弱点。因此,为了加强集群的分离,我们应用后端规范,在层次的每个层次对节点施加最大间断限制。在本文中,我们说明框架如何与BHMC实现原始的改进。