Most existing methods in vision-language retrieval match two modalities by either comparing their global feature vectors which misses sufficient information and lacks interpretability, detecting objects in images or videos and aligning the text with fine-grained features which relies on complicated model designs, or modeling fine-grained interaction via cross-attention upon visual and textual tokens which suffers from inferior efficiency. To address these limitations, some recent works simply aggregate the token-wise similarities to achieve fine-grained alignment, but they lack intuitive explanations as well as neglect the relationships between token-level features and global representations with high-level semantics. In this work, we rethink fine-grained cross-modal alignment and devise a new model-agnostic formulation for it. We additionally demystify the recent popular works and subsume them into our scheme. Furthermore, inspired by optimal transport theory, we introduce TokenFlow, an instantiation of the proposed scheme. By modifying only the similarity function, the performance of our method is comparable to the SoTA algorithms with heavy model designs on major video-text retrieval benchmarks. The visualization further indicates that TokenFlow successfully leverages the fine-grained information and achieves better interpretability.
翻译:视觉-语言检索的大多数现有方法都与两种模式相匹配:要么比较缺乏足够信息且缺乏解释性的全球特征矢量,在图像或视频中探测对象,使文本与依赖复杂模型设计的精细雕刻特征相匹配,要么通过对低效率的视觉和文字象征的交叉关注进行微细的模拟互动。为了解决这些局限性,最近的一些工作只是汇总了象征性的相似之处,以达到细微的调整,但是它们缺乏直观的解释,并且忽视了象征性特征与高层次语义学的全球表现之间的关系。在这项工作中,我们重新思考了精细的跨模式对齐,并设计了一种新的模型-认知的配方。我们进一步解开最近流行的作品,并将其纳入我们的方案。此外,在最佳运输理论的启发下,我们引入了TokenFlowlow,这是拟议的办法的即即即即即是。通过只修改相似性功能,我们的方法的性能与SolowTA与主要视频-文字检索基准的重模型设计相似。视觉-视觉化进一步表明Token-Fregality 能够成功地理解。