Traditional clustering algorithms such as K-means rely heavily on the nature of the chosen metric or data representation. To get meaningful clusters, these representations need to be tailored to the downstream task (e.g. cluster photos by object category, cluster faces by identity). Therefore, we frame clustering as a meta-learning task, few-shot clustering, which allows us to specify how to cluster the data at the meta-training level, despite the clustering algorithm itself being unsupervised. We propose Centroid Networks, a simple and efficient few-shot clustering method based on learning representations which are tailored both to the task to solve and to its internal clustering module. We also introduce unsupervised few-shot classification, which is conceptually similar to few-shot clustering, but is strictly harder than supervised* few-shot classification and therefore allows direct comparison with existing supervised few-shot classification methods. On Omniglot and miniImageNet, our method achieves accuracy competitive with popular supervised few-shot classification algorithms, despite using *no labels* from the support set. We also show performance competitive with state-of-the-art learning-to-cluster methods.
翻译:K手段等传统集群算法在很大程度上依赖所选择的计量或数据代表的性质。为了获得有意义的分类组合,这些表达方式需要适应下游任务(例如按对象类别分类的群集照片,按身份分类的群集面),因此,我们把群集作为一种元学习任务、少发集成,使我们能够具体说明如何在元培训一级将数据分组,尽管群集算法本身不受监督。我们提议了中网络,这是一种简单而高效的、以学习表达方式为基础、以学习表达方式为基础、既适合解决任务又适合内部群集模块的少发集集集成法为基础的简单、少发集集成法。我们还采用了非监督的少发分类法,这在概念上与少数发集集群相似,但严格地比受监督的分类法更难。在Omniglot和MiniImaget上,我们的方法与受监督的受公众监督的少发分类算法实现了准确性竞争,尽管使用了支助集的* no lables* 。我们还展示了与最新学习组方法的竞争力。