The past two decades have seen increasingly rapid advances in the field of multi-view representation learning due to it extracting useful information from diverse domains to facilitate the development of multi-view applications. However, the community faces two challenges: i) how to learn robust representations from a large amount of unlabeled data to against noise or incomplete views setting, and ii) how to balance view consistency and complementary for various downstream tasks. To this end, we utilize a deep fusion network to fuse view-specific representations into the view-common representation, extracting high-level semantics for obtaining robust representation. In addition, we employ a clustering task to guide the fusion network to prevent it from leading to trivial solutions. For balancing consistency and complementary, then, we design an asymmetrical contrastive strategy that aligns the view-common representation and each view-specific representation. These modules are incorporated into a unified method known as CLustering-guided cOntrastiVE fusioN (CLOVEN). We quantitatively and qualitatively evaluate the proposed method on five datasets, demonstrating that CLOVEN outperforms 11 competitive multi-view learning methods in clustering and classification. In the incomplete view scenario, our proposed method resists noise interference better than those of our competitors. Furthermore, the visualization analysis shows that CLOVEN can preserve the intrinsic structure of view-specific representation while also improving the compactness of view-commom representation. Our source code will be available soon at https://github.com/guanzhou-ke/cloven.
翻译:过去二十年来,在多视角代表性学习领域取得了日益迅速的进展,因为它从不同领域提取了有用的信息,以促进多视角应用程序的开发。然而,社区面临两个挑战:(1) 如何从大量未贴标签的数据中学习强有力的代表性,以对抗噪音或不完整的视角设置,(2) 如何平衡地看待各种下游任务的一致性和互补性。为此,我们利用一个深层融合网络,将特定视角的表述融入视野-共同代表性,为获得强有力的代表性提取高层次的语义学。此外,我们利用一个集群任务来指导聚合网络,防止其导致微不足道的解决方案。为了平衡一致性和互补性,我们设计了一个对称对称对比对比的对比战略,使视觉-共同代表性和每种特定视角的表述相一致,这些模块被纳入一个称为CLustering-制导的Onttrasteve fusioN(CLOVEN)的统一方法。我们将对五套数据集的拟议方法进行定量和定性评估,表明,CLVE/CLVI将很快超越具有竞争力的多视角的多视角学习观点,同时显示我们所拟议的CR Revormission-view viewsal roup laus thesal disal laveal