Expectation maximisation (EM) is usually thought of as an unsupervised learning method for estimating the parameters of a mixture distribution, however it can also be used for supervised learning when class labels are available. As such, EM has been applied to train neural nets including the probabilistic radial basis function (PRBF) network or shared kernel (SK) model. This paper addresses two major shortcomings of previous work in this area: the lack of rigour in the derivation of the EM training algorithm; and the computational complexity of the technique, which has limited it to low dimensional data sets. We first present a detailed derivation of EM for the Gaussian shared kernel model PRBF classifier, making use of data association theory to obtain the complete data likelihood, Baum's auxiliary function (the E-step) and its subsequent maximisation (M-step). To reduce complexity of the resulting SKEM algorithm, we partition the feature space into $R$ non-overlapping subsets of variables. The resulting product decomposition of the joint data likelihood, which is exact when the feature partitions are independent, allows the SKEM to be implemented in parallel and at $R^2$ times lower complexity. The operation of the partitioned SKEM algorithm is demonstrated on the MNIST data set and compared with its non-partitioned counterpart. It eventuates that improved performance at reduced complexity is achievable. Comparisons with standard classification algorithms are provided on a number of other benchmark data sets.
翻译:期望最大化(EM)通常被认为是估算混合物分布参数的一种不受监督的学习方法,但也可以在有类标签时用于监督学习。因此,EM被用于培训神经网,包括概率性半径基功能(PRBF)网络或共享内核(SK)模型。本文述及这一领域先前工作的两个重大缺陷:EM培训算法的衍生缺乏严格性;以及技术的计算复杂性,这限制了该技术的低维数据集。我们首先为Gaussian 共享内核模型PRBF分类仪详细介绍了EM的衍生,利用数据关联理论来获得完整的数据可能性,Baum的辅助功能(E级)及其随后的最大化(M级)。为降低SKEM算法的复杂性,我们将特性空间分成非重叠的变量组。因此,将联合数据的可能性进行分解,当Gaussian 共享内核模型PRBFF分类仪的分类仪模型时,我们首先详细介绍了EM-R值的详细推算结果,当特性分解率比值较低的SMQ标准数时,使得SEM值的平行的运行到S-rlalalalalalalalalation 。 在S-dealalalalalalalalalation 上,在Slades上提供S-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx。