With the development of depth sensors in recent years, RGBD object tracking has received significant attention. Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference. However, some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored. On the other hand, some methods attempt to fuse the two modalities by treating them equally, resulting in the missing of modality-specific features. To tackle these limitations, we propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking. The first fusion module focuses on extracting the shared information between modalities based on cross-modal attention. The second aims at integrating the RGB-specific and depth-specific information to enhance the fused features. By fusing both the modality-shared and modality-specific information in a modality-aware scheme, our DMTracker can learn discriminative representations in complex tracking scenes. Experiments show that our proposed tracker achieves very promising results on challenging RGBD benchmarks.
翻译:近年来,随着深度传感器的发展,RGBD物体跟踪受到高度重视。与传统的RGB物体跟踪相比,增加深度模式可以有效解决目标和背景干扰问题。然而,一些现有的RGBD追踪器分别使用这两种模式,因此它们之间分享的一些特别有用的信息被忽视。另一方面,有些方法试图通过同等对待这两种模式,从而整合这两种模式,从而导致模式特性的缺失。为克服这些限制,我们提议采用新的双刃双刃混合模式跟踪器(MDMTTracker),目的是了解目标目标物体的信息化和有区别的表达方式,以便进行强有力的RGBD跟踪。第一个聚变模块侧重于在基于跨模式关注的模式之间获取共享的信息。第二个模块旨在将RGB特定和深度特定信息整合起来,以加强集成特征。通过在模式认知计划中同时使用模式共享的信息和特定模式信息,我们的DMTracker可以在复杂的跟踪场上学习有区别的描述。实验显示,我们提议的跟踪器在挑战的 RGBD基准上取得了很有希望的结果。