Multi-View Clustering (MVC) has garnered increasing attention in recent years. It is capable of partitioning data samples into distinct groups by learning a consensus representation. However, a significant challenge remains: the problem of untrustworthy fusion. This problem primarily arises from two key factors: 1) Existing methods often ignore the presence of inherent noise within individual views; 2) In traditional MVC methods using Contrastive Learning (CL), similarity computations typically rely on different views of the same instance, while neglecting the structural information from nearest neighbors within the same cluster. Consequently, this leads to the wrong direction for multi-view fusion. To address this problem, we present a novel Trusted Hierarchical Contrastive Representation Learning (THCRL). It consists of two key modules. Specifically, we propose the Deep Symmetry Hierarchical Fusion (DSHF) module, which leverages the UNet architecture integrated with multiple denoising mechanisms to achieve trustworthy fusion of multi-view data. Furthermore, we present the Average K-Nearest Neighbors Contrastive Learning (AKCL) module to align the fused representation with the view-specific representation. Unlike conventional strategies, AKCL enhances representation similarity among samples belonging to the same cluster, rather than merely focusing on the same sample across views, thereby reinforcing the confidence of the fused representation. Extensive experiments demonstrate that THCRL achieves the state-of-the-art performance in deep MVC tasks.
翻译:近年来,多视图聚类(MVC)受到越来越多的关注。它能够通过学习一致性表征将数据样本划分为不同的组。然而,一个重要的挑战依然存在:不可信融合问题。该问题主要源于两个关键因素:1)现有方法往往忽略单个视图中存在的固有噪声;2)在使用对比学习(CL)的传统MVC方法中,相似性计算通常依赖于同一实例的不同视图,而忽视了同一聚类内最近邻的结构信息。因此,这导致多视图融合的方向出现偏差。为解决此问题,我们提出了一种新颖的可信分层对比表征学习方法(THCRL)。它包含两个关键模块。具体而言,我们提出了深度对称分层融合(DSHF)模块,该模块利用集成多种去噪机制的UNet架构,实现多视图数据的可信融合。此外,我们提出了平均K近邻对比学习(AKCL)模块,用于将融合表征与视图特定表征对齐。与常规策略不同,AKCL增强了属于同一聚类的样本之间的表征相似性,而非仅仅关注跨视图的同一样本,从而强化了融合表征的可信度。大量实验表明,THCRL在深度MVC任务中实现了最先进的性能。