Video advertisement content structuring aims to segment a given video advertisement and label each segment on various dimensions, such as presentation form, scene, and style. Different from real-life videos, video advertisements contain sufficient and useful multi-modal content like caption and speech, which provides crucial video semantics and would enhance the structuring process. In this paper, we propose a multi-modal encoder to learn multi-modal representation from video advertisements by interacting between video-audio and text. Based on multi-modal representation, we then apply Boundary-Matching Network to generate temporal proposals. To make the proposals more accurate, we refine generated proposals by scene-guided alignment and re-ranking. Finally, we incorporate proposal located embeddings into the introduced multi-modal encoder to capture temporal relationships between local features of each proposal and global features of the whole video for classification. Experimental results show that our method achieves significantly improvement compared with several baselines and Rank 1 on the task of Multi-modal Ads Video Understanding in ACM Multimedia 2021 Grand Challenge. Ablation study further shows that leveraging multi-modal content like caption and speech in video advertisements significantly improve the performance.
翻译:视频内容结构的视频广告内容结构旨在将特定视频广告和每个部分的标签分为不同层面,如演示形式、场景和风格。与真实存在的视频不同,视频广告包含充分和有用的多模式内容,如字幕和演讲,提供关键的视频语义,并将加强结构过程。在本文中,我们提议一个多模式编码器,通过视频视频视频和文字之间的互动,从视频广告中学习多模式的表述。基于多模式代表制,我们然后应用边界匹配网络生成时间建议。为了使建议更加准确,我们通过场景引导校正和重新排序改进了建议。最后,我们纳入了在引入的多模式编码器中嵌入位置的建议,以捕捉每个提案的本地特征与整个视频分类的全球特征之间的时间关系。实验结果表明,我们的方法与若干基线相比有了很大的改进,与2021年ACM多媒体多媒体多模式视频理解任务第1级相比,取得了显著的改进。进一步显示,在视频广告中利用多模式内容(如字幕和语音)大大改进了业绩。