The task of image-text matching aims to map representations from different modalities into a common joint visual-textual embedding. However, the most widely used datasets for this task, MSCOCO and Flickr30K, are actually image captioning datasets that offer a very limited set of relationships between images and sentences in their ground-truth annotations. This limited ground truth information forces us to use evaluation metrics based on binary relevance: given a sentence query we consider only one image as relevant. However, many other relevant images or captions may be present in the dataset. In this work, we propose two metrics that evaluate the degree of semantic relevance of retrieved items, independently of their annotated binary relevance. Additionally, we incorporate a novel strategy that uses an image captioning metric, CIDEr, to define a Semantic Adaptive Margin (SAM) to be optimized in a standard triplet loss. By incorporating our formulation to existing models, a \emph{large} improvement is obtained in scenarios where available training data is limited. We also demonstrate that the performance on the annotated image-caption pairs is maintained while improving on other non-annotated relevant items when employing the full training set. Code with our metrics and adaptive margin formulation will be made public.
翻译:图像- 文本匹配任务旨在将不同模式的表达方式映射成共同的视觉- 文字嵌入式。 然而, 最广泛用于此任务的数据集( MSCOCO 和 Flickr30K ) 实际上是图像字幕数据集, 在图像和判决的地面真相说明中提供了非常有限的一组关系。 这种有限的地面真相信息迫使我们使用基于二元相关性的评价指标: 给一个句子查询, 我们认为只有一个图像是相关的。 但是, 数据集中可能存在许多其他相关的图像或说明。 在这项工作中, 我们提出了两个衡量标准, 评估检索到的物品的语义相关性程度, 独立于其附加说明的二元相关性。 此外, 我们纳入了一个新的战略, 使用图像说明性说明性指标( CIDer) 来定义一个基于标准的三重损失的语义调整边距值( SAM ) 。 通过将我们的配方与现有模型相结合, 在现有的培训数据有限的情况下, 也取得了一项改进。 我们还表明, 附加说明性图像描述性图像组合的性能保持完整调整性, 同时将使用其它的校准度设定非标准。