When a recurrent neural network language model is used for caption generation, the image information can be fed to the neural network either by directly incorporating it in the RNN -- conditioning the language model by `injecting' image features -- or in a layer following the RNN -- conditioning the language model by `merging' image features. While both options are attested in the literature, there is as yet no systematic comparison between the two. In this paper we empirically show that it is not especially detrimental to performance whether one architecture is used or another. The merge architecture does have practical advantages, as conditioning by merging allows the RNN's hidden state vector to shrink in size by up to four times. Our results suggest that the visual and linguistic modalities for caption generation need not be jointly encoded by the RNN as that yields large, memory-intensive models with few tangible advantages in performance; rather, the multimodal integration should be delayed to a subsequent stage.
翻译:当一个经常性神经网络语言模型被用于制作标题时,图像信息可以直接将其纳入神经网络,办法是将它纳入RNN -- -- 将语言模型设置为“输入”图像特征,或者在RNN之后的一层 -- -- 将语言模型设置为“合并”图像特征。虽然这两个选项在文献中得到证明,但两者之间还没有系统性的比较。在本文中,我们从经验上表明,无论使用一个架构还是另一个架构都不特别有害于性能。合并结构确实具有实际优势,因为通过合并调整,RNN隐藏的状态矢量的缩小到四倍。我们的结果表明,标题生成的视觉和语言模式不需要由RNN联合编码,因为这样产生大型、记忆密集的模型,在性能上没有多少明显优势;相反,多式集成应推迟到下一个阶段。