As one of the most fundamental techniques in multimodal learning, cross-modal matching aims to project various sensory modalities into a shared feature space. To achieve this, massive and correctly aligned data pairs are required for model training. However, unlike unimodal datasets, multimodal datasets are extremely harder to collect and annotate precisely. As an alternative, the co-occurred data pairs (e.g., image-text pairs) collected from the Internet have been widely exploited in the area. Unfortunately, the cheaply collected dataset unavoidably contains many mismatched data pairs, which have been proven to be harmful to the model's performance. To address this, we propose a general framework called BiCro (Bidirectional Cross-modal similarity consistency), which can be easily integrated into existing cross-modal matching models and improve their robustness against noisy data. Specifically, BiCro aims to estimate soft labels for noisy data pairs to reflect their true correspondence degree. The basic idea of BiCro is motivated by that -- taking image-text matching as an example -- similar images should have similar textual descriptions and vice versa. Then the consistency of these two similarities can be recast as the estimated soft labels to train the matching model. The experiments on three popular cross-modal matching datasets demonstrate that our method significantly improves the noise-robustness of various matching models, and surpass the state-of-the-art by a clear margin.
翻译:作为多模态学习中最基本的技术之一,跨模态匹配旨在将各种感觉模态投影到共享特征空间中。为了实现这一目标,需要大量正确对齐的数据对进行模型训练。然而,与单模态数据集不同,多模态数据集极其难以精确地收集和注释。作为替代,从互联网收集到的共现数据对(例如图像-文本对)已被广泛地用于此领域。不幸的是,便宜收集的数据集不可避免地包含许多不匹配的数据对,这已被证明会对模型的性能造成不利影响。为了解决这个问题,我们提出了一个通用框架,称为BiCro(双向跨模态相似度一致性),可以轻松地集成到现有的跨模态匹配模型中,并提高其对噪声数据的鲁棒性。具体而言,BiCro旨在为噪声数据对估计软标签,以反映它们的真实对应度。 BiCro的基本思想是:以图像-文本匹配为例,相似的图像应具有相似的文本描述,反之亦然。然后,可以重构这两个相似性的一致性,作为训练匹配模型的估计软标签。在三个流行的跨模态匹配数据集实验中,我们的方法显著提高了各种匹配模型的噪声鲁棒性,并超过了最新技术水平。