Long-range and short-range temporal modeling are two complementary and crucial aspects of video recognition. Most of the state-of-the-arts focus on short-range spatio-temporal modeling and then average multiple snippet-level predictions to yield the final video-level prediction. Thus, their video-level prediction does not consider spatio-temporal features of how video evolves along the temporal dimension. In this paper, we introduce a novel Dynamic Segment Aggregation (DSA) module to capture relationship among snippets. To be more specific, we attempt to generate a dynamic kernel for a convolutional operation to aggregate long-range temporal information among adjacent snippets adaptively. The DSA module is an efficient plug-and-play module and can be combined with the off-the-shelf clip-based models (i.e., TSM, I3D) to perform powerful long-range modeling with minimal overhead. The final video architecture, coined as DSANet. We conduct extensive experiments on several video recognition benchmarks (i.e., Mini-Kinetics-200, Kinetics-400, Something-Something V1 and ActivityNet) to show its superiority. Our proposed DSA module is shown to benefit various video recognition models significantly. For example, equipped with DSA modules, the top-1 accuracy of I3D ResNet-50 is improved from 74.9% to 78.2% on Kinetics-400. Codes are available at https://github.com/whwu95/DSANet.
翻译:长程和短程时间建模是视频识别的两个互补和关键方面。大多数最先进的艺术侧重于短程时空建模,然后是平均多片级预测,以得出最后视频级的预测。因此,其视频级预测不考虑视频在时间维度上如何演进的片段时空特征。在本文件中,我们引入了一个新型的动态部分聚合模块(DSA)以捕捉片片间的关系。更具体地说,我们试图生成一个动态内核,用于在相邻的狙击场间进行动态操作,以汇总长程长时间建模。DSA模块是一个高效的插接和播放模块,可以与视频外机基模型(即D.e.,TSM,I3D)结合,以最小的间接模式进行强有力的远程建模。最后的视频结构,以DSANet的形式生成。我们在多个视频识别基准上进行广泛的实验(i.e,Mini-Kinetits-200,Kintical-Net1-400, KiniticalSA-400) 模块展示了其高级的V.