We propose a segmentation-based bounding box generation method for omnidirectional pedestrian detection that enables detectors to tightly fit bounding boxes to pedestrians without omnidirectional images for training. Due to the wide angle of view, omnidirectional cameras are more cost-effective than standard cameras and hence suitable for large-scale monitoring. The problem of using omnidirectional cameras for pedestrian detection is that the performance of standard pedestrian detectors is likely to be substantially degraded because pedestrians' appearance in omnidirectional images may be rotated to any angle. Existing methods mitigate this issue by transforming images during inference. However, the transformation substantially degrades the detection accuracy and speed. A recently proposed method obviates the transformation by training detectors with omnidirectional images, which instead incurs huge annotation costs. To obviate both the transformation and annotation works, we leverage an existing large-scale object detection dataset. We train a detector with rotated images and tightly fitted bounding box annotations generated from the segmentation annotations in the dataset, resulting in detecting pedestrians in omnidirectional images with tightly fitted bounding boxes. We also develop pseudo-fisheye distortion augmentation, which further enhances the performance. Extensive analysis shows that our detector successfully fits bounding boxes to pedestrians and demonstrates substantial performance improvement.
翻译:我们建议一种基于分解的捆绑箱生成方法,用于进行全向行人检测,使探测器能够在没有全向图像的情况下,使行人与行人紧紧捆绑盒子,不需接受全向式培训的图像。由于视野宽广,全向照相机比标准相机更具成本效益,因此适合大规模监测。使用全向式照相机进行行人检测的问题是,标准行人探测器的性能可能会大大降低,因为行人出现在全向式图像中的性能可能被旋转到任何角度。现有的方法通过在感知期间转换图像来缓解这一问题。然而,这种转换会大大降低检测的准确性和速度。最近提出的一种方法避免了使用全向式图像培训探测器的转换,而这种转换需要巨大的注解成本。为避免转换和批注工作,我们利用现有的大型天体探测器的性探测数据集。我们用旋转的图像来训练一台探测器,并用从数据设置的分解图解中生成的紧凑合的带框条纹图解,从而在感测中检测行人的准确度和速度的精确性变变,从而显示我们的模拟图像。