Zero-Shot Learning (ZSL) in video classification is a promising research direction, which aims to tackle the challenge from explosive growth of video categories. Most existing methods exploit seen-to-unseen correlation via learning a projection between visual and semantic spaces. However, such projection-based paradigms cannot fully utilize the discriminative information implied in data distribution, and commonly suffer from the information degradation issue caused by "heterogeneity gap". In this paper, we propose a visual data synthesis framework via GAN to address these problems. Specifically, both semantic knowledge and visual distribution are leveraged to synthesize video feature of unseen categories, and ZSL can be turned into typical supervised problem with the synthetic features. First, we propose multi-level semantic inference to boost video feature synthesis, which captures the discriminative information implied in joint visual-semantic distribution via feature-level and label-level semantic inference. Second, we propose Matching-aware Mutual Information Correlation to overcome information degradation issue, which captures seen-to-unseen correlation in matched and mismatched visual-semantic pairs by mutual information, providing the zero-shot synthesis procedure with robust guidance signals. Experimental results on four video datasets demonstrate that our approach can improve the zero-shot video classification performance significantly.
翻译:视频分类中的零热学习(ZSL)是一个很有希望的研究方向,目的是应对视频类别爆炸性增长带来的挑战。大多数现有方法都利用视觉和语义空间之间的预测,利用视觉和语义空间之间的可见到不见的关联。然而,这种基于预测的范例不能充分利用数据分布中隐含的歧视性信息,而且通常会因“异质性差”造成的信息退化问题而受害。在本文件中,我们提议通过GAN建立一个视觉数据合成框架,以解决这些问题。具体地说,语义知识和视觉传播都被用来合成看不见类别视频特征的合成特征,而ZSL可以变成典型的合成特征监督问题。首先,我们提出多层次的语义推论,以强化视频特征合成,从而捕捉通过地貌级别和标签等级的语义性格推断联合视觉到的歧视性信息。第二,我们提议通过匹配认知的相互信息关联性信息关联性信息来克服信息退化问题,通过匹配和不相匹配的视觉和不匹配的视觉和不吻合的图像模拟实验模型,通过相互信息来显著地展示可靠的合成结果。