The pioneering method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling. This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data. However, it often suffers from label inconsistency or limited diversity, which leads to poor performance. In this work, we prove that the core reason for this is lack of a clustering-friendly property in the embedding space. We address this by minimizing the inter- to intra-class similarity ratio to provide clustering-friendly embedding features, and validate our approach through comprehensive experiments. Note that, despite only utilizing a simple clustering algorithm (k-means) in our embedding space to obtain the pseudo-labels, we achieve significant improvement. Moreover, we adopt a progressive evaluation mechanism to obtain more diverse samples in order to further alleviate the limited diversity problem. Finally, our approach is also model-agnostic and can easily be integrated into existing supervised methods. To demonstrate its generalization ability, we integrate it into two representative algorithms: MAML and EP. The results on three main few-shot benchmarks clearly show that the proposed method achieves significant improvement compared to state-of-the-art models. Notably, our approach also outperforms the corresponding supervised method in two tasks.
翻译:无监督的元学习的开创性方法CACTUs(CACTUs)是使用假标签的基于集群的方法。 这种方法是模型的不可知性,可以与受监督的算法相结合,从未贴标签的数据中学习。 但是,它往往受到标签不一致或多样性有限的影响,这导致工作表现不佳。 在这项工作中,我们证明,造成这种情况的核心原因是在嵌入空间缺乏有利于集群的属性。 我们通过尽可能减少类内相似性比率来解决这个问题,以提供有利于集群的嵌入功能,并通过全面试验来验证我们的方法。 注意,尽管在嵌入空间中只使用简单的集群算法(k-pools)来获取假标签,但我们还是取得了显著的改进。 此外,我们采用了一种渐进的评价机制,以获得更多样化的样本,从而进一步缓解有限的多样性问题。 最后,我们的方法也是模型-不可忽视的,可以很容易地融入现有的监督方法。 为了证明其普及性,我们将其纳入两种具有代表性的算法:MAML和EP。 注意,尽管在三个主要的少数基准上的结果清楚地表明,我们所建议的方法在两个不同的模式中也明显地显示了我们所监督的改进的方法。