The instability is shown in the existing methods of representation learning based on Euclidean distance under a broad set of conditions. Furthermore, the scarcity and high cost of labels prompt us to explore more expressive representation learning methods which depends on the labels as few as possible. To address these issues, the small-perturbation ideology is firstly introduced on the representation learning model based on the representation probability distribution. The positive small-perturbation information (SPI) which only depend on two labels of each cluster is used to stimulate the representation probability distribution and then two variant models are proposed to fine-tune the expected representation distribution of RBM, namely, Micro-supervised Disturbance GRBM (Micro-DGRBM) and Micro-supervised Disturbance RBM (Micro-DRBM) models. The Kullback-Leibler (KL) divergence of SPI is minimized in the same cluster to promote the representation probability distributions to become more similar in Contrastive Divergence(CD) learning. In contrast, the KL divergence of SPI is maximized in the different clusters to enforce the representation probability distributions to become more dissimilar in CD learning. To explore the representation learning capability under the continuous stimulation of the SPI, we present a deep Micro-supervised Disturbance Learning (Micro-DL) framework based on the Micro-DGRBM and Micro-DRBM models and compare it with a similar deep structure which has not any external stimulation. Experimental results demonstrate that the proposed deep Micro-DL architecture shows better performance in comparison to the baseline method, the most related shallow models and deep frameworks for clustering.
翻译:在一系列广泛条件下,基于Euclidean距离的现有代表制学习方法显示不稳定。此外,标签稀缺且成本高,促使我们探索更直观的代议制学习方法,这种方法取决于尽可能少的标签。为了解决这些问题,根据代表概率分布的代议制学习模式首先引入了小扰动意识形态。仅依赖每个组的两个标签的小型扰动信息(SPI)用于刺激代表概率分布,然后提出两个深层次的缩微模型来微调成果管理制的预期代表性分布,即Micro Superousalalcurate GRBM(Micro-DGRBM)和Micmicrovisurate DRMMBM(Mcro-DRBM)模式。在同一组中,Spreback-Leber(KL)差异最小化信息(KL)最小化信息,以促进代表概率分布在对比性DVI(CD)中更加相似的代议式模型。相比之下,SPI(KL)差异差异是微缩缩缩缩缩缩缩缩缩缩图框架框架,在不断学习SBRBIBRMBI的外部分析框架中,在不断的缩缩缩缩缩缩分析模型分析模型中,在深度分析模型中,在深度分析模型中,在深度分析模型中,在不断学习中,在不断学习。