We introduce a new approach to probabilistic unsupervised learning based on the recognition-parametrised model (RPM): a normalised semi-parametric hypothesis class for joint distributions over observed and latent variables. Under the key assumption that observations are conditionally independent given the latents, RPMs directly encode the "recognition" process, parametrising both the prior distribution on the latents and their conditional distributions given observations. This recognition model is paired with non-parametric descriptions of the marginal distribution of each observed variable. Thus, the focus is on learning a good latent representation that captures dependence between the measurements. The RPM permits exact maximum likelihood learning in settings with discrete latents and a tractable prior, even when the mapping between continuous observations and the latents is expressed through a flexible model such as a neural network. We develop effective approximations for the case of continuous latent variables with tractable priors. Unlike the approximations necessary in dual-parametrised models such as Helmholtz machines and variational autoencoders, these RPM approximations introduce only minor bias, which may often vanish asymptotically. Furthermore, where the prior on latents is intractable the RPM may be combined effectively with standard probabilistic techniques such as variational Bayes. We demonstrate the model in high dimensional data settings, including a form of weakly supervised learning on MNIST digits and the discovery of latent maps from sensory observations. The RPM provides an effective way to discover, represent and reason probabilistically about the latent structure underlying observational data, functions which are critical to both animal and artificial intelligence.
翻译:我们引入了一种基于认知偏差模型(RPM)的概率性非监督性学习的新方法:一个正常的半参数假假设类,用于对观测和潜伏变量进行联合分布。在关键假设中,观测是有条件独立的,因为潜伏,RPM直接编码了“识别”过程,将先前对潜伏的分布及其给定的有条件分布相形见绌。这个识别模型与对观察到的每个变量的边际分布的非参数描述相配。因此,重点是学习一种良好的潜在代表,可以捕捉测量之间的依赖性。RMMP允许在具有离散潜伏和可移动前的环境下进行精确的潜值结构学习,即使连续观测和潜伏之间的映像通过神经网络等灵活的模型来表达。我们为持续潜伏变量与可移动的前几点观测。与Helmhotz 机器和变形自动解码模型中所需的近似值不同,这些RPMPM只能引入轻微的直径偏差,这些暗度结构只能代表前期的直径直径直径的直径,而前方的直径的直径直径直径直径直径直径,从而可以展示数据在前方的直径向中学习。