In this work we present two video test data sets for the novel computer vision (CV) task of out of distribution tracking (OOD tracking). Here, OOD objects are understood as objects with a semantic class outside the semantic space of an underlying image segmentation algorithm, or an instance within the semantic space which however looks decisively different from the instances contained in the training data. OOD objects occurring on video sequences should be detected on single frames as early as possible and tracked over their time of appearance as long as possible. During the time of appearance, they should be segmented as precisely as possible. We present the SOS data set containing 20 video sequences of street scenes and more than 1000 labeled frames with up to two OOD objects. We furthermore publish the synthetic CARLA-WildLife data set that consists of 26 video sequences containing up to four OOD objects on a single frame. We propose metrics to measure the success of OOD tracking and develop a baseline algorithm that efficiently tracks the OOD objects. As an application that benefits from OOD tracking, we retrieve OOD sequences from unlabeled videos of street scenes containing OOD objects.
翻译:在这项工作中,我们为新计算机视觉的分发跟踪任务(OOD跟踪)提供了两套视频测试数据集。在这里,OOD物体被理解为在基本图像分解算法的语义空间外带有语义类的物体,或者在语义空间内带有一种但与培训数据中包含的情况有决定性差别的示例。视频序列上出现的OOD物体应尽早在单一框上检测,并尽可能长的时间在外观时跟踪。在外观期间,应尽可能精确地对它们进行分解。我们提供SOS数据集,其中包含20个街头场景的视频序列和1 000多个贴有标签的框,多达两个OOD物体。我们还出版了合成的CARLA-WildLife数据集,由26个视频序列组成,在单一框内包含多达4个OOD物体的视频组成。我们提出了衡量OD跟踪成功与否的衡量标准,并开发一个有效跟踪OOD物体的基线算法。作为OOD追踪的一个应用,我们从含有OOD物体的街道未贴标签的视频中检索OD的OD序列。