Recent works have shown huge success of deep learning models for common in vocabulary (IV) scene text recognition. However, in real-world scenarios, out-of-vocabulary (OOV) words are of great importance and SOTA recognition models usually perform poorly on OOV settings. Inspired by the intuition that the learned language prior have limited OOV preformence, we design a framework named Vision Language Adaptive Mutual Decoder (VLAMD) to tackle OOV problems partly. VLAMD consists of three main conponents. Firstly, we build an attention based LSTM decoder with two adaptively merged visual-only modules, yields a vision-language balanced main branch. Secondly, we add an auxiliary query based autoregressive transformer decoding head for common visual and language prior representation learning. Finally, we couple these two designs with bidirectional training for more diverse language modeling, and do mutual sequential decoding to get robuster results. Our approach achieved 70.31\% and 59.61\% word accuracy on IV+OOV and OOV settings respectively on Cropped Word Recognition Task of OOV-ST Challenge at ECCV 2022 TiE Workshop, where we got 1st place on both settings.
翻译:最近的工作显示,在词汇(IV)现场文字识别中共同的深层次学习模式取得了巨大成功;然而,在现实世界情景中,校外字非常重要,SOTA识别模式在OOV设置中通常效果不佳。受以下直觉的启发,即以前学习的语言对OOOV的预发型作用有限,我们设计了一个称为“视觉语言适应性适应性兼容共解码器(VLAMD)”的框架,以解决OOOV问题。VLAMD由三个主要响应方组成。首先,我们用两个适应性合并的视觉单一模块建立一个基于LSTM解码器的注意点,产生一个视觉平衡的主要分支。第二,我们增加了一个基于自动递增变异变变器在通用视觉和语言之前的演示学习中解码头的辅助查询。最后,我们将这两个设计与关于更多样化语言建模的双向培训结合起来,并进行相继解码以获得更稳健的结果。我们的方法在IV+OV和OVE设置中实现了70.31 ⁇ 59.61字精准,产生视觉平衡。我们分别在EC Ti22 TI22 的作物识别识别任务中,我们在这里都获得了了OV.