The canonical approach to video-and-language learning (e.g., video question answering) dictates a neural model to learn from offline-extracted dense video features from vision models and text features from language models. These feature extractors are trained independently and usually on tasks different from the target domains, rendering these fixed features sub-optimal for downstream tasks. Moreover, due to the high computational overload of dense video features, it is often difficult (or infeasible) to plug feature extractors directly into existing approaches for easy finetuning. To provide a remedy to this dilemma, we propose a generic framework ClipBERT that enables affordable end-to-end learning for video-and-language tasks, by employing sparse sampling, where only a single or a few sparsely sampled short clips from a video are used at each training step. Experiments on text-to-video retrieval and video question answering on six datasets demonstrate that ClipBERT outperforms (or is on par with) existing methods that exploit full-length videos, suggesting that end-to-end learning with just a few sparsely sampled clips is often more accurate than using densely extracted offline features from full-length videos, proving the proverbial less-is-more principle. Videos in the datasets are from considerably different domains and lengths, ranging from 3-second generic domain GIF videos to 180-second YouTube human activity videos, showing the generalization ability of our approach. Comprehensive ablation studies and thorough analyses are provided to dissect what factors lead to this success. Our code is publicly available at https://github.com/jayleicn/ClipBERT
翻译:视频和语言学习(例如,视频问题解答)的常规方法要求一种神经模型,以便从视觉模型和语言模型的文字特征中从离线的密集视频特征中学习。这些特征提取器是独立培训的,通常用于与目标领域不同的任务,使这些固定特征在下游任务中具有亚最佳性。此外,由于密集视频特征的计算超载率很高,将特征提取器直接插入现有方法以方便微调。为了纠正这一困境,我们建议了一个通用框架 ClipBERT,通过使用稀有的取样,使视频和语言任务能够以可承受的直线式直线直线视频分析。每个培训步骤只使用一个或几个分散抽样的短短片来完成这些任务,使这些固定特征在下游任务中达到亚性最优化。在6个数据集上对文本到视频检索和视频问题的实验表明,ClipBERT超越(或接近于)现有方法,从而导致全程视频的全程。我们从一个尾端到终端的视频,表明仅用少量的全程能力对域域域域域分析,从一个小的视频中,从一个普通的视频活动到从一个普通视频流到另一个视频流到另一个的视频,通常的视频,通常的视频是更不甚甚甚甚深的视频,从一个从一个比高的图像的图像中,而从整个的图像的图像的视频,从一个更精确的视频,从一个图像到更精确的图像,常常地段,常常地段,从一个从整个的视频是提供。