Detection of small objects and objects far away in the scene is a major challenge in surveillance applications. Such objects are represented by small number of pixels in the image and lack sufficient details, making them difficult to detect using conventional detectors. In this work, an open-source framework called Slicing Aided Hyper Inference (SAHI) is proposed that provides a generic slicing aided inference and fine-tuning pipeline for small object detection. The proposed technique is generic in the sense that it can be applied on top of any available object detector without any fine-tuning. Experimental evaluations, using object detection baselines on the Visdrone and xView aerial object detection datasets show that the proposed inference method can increase object detection AP by 6.8%, 5.1% and 5.3% for FCOS, VFNet and TOOD detectors, respectively. Moreover, the detection accuracy can be further increased with a slicing aided fine-tuning, resulting in a cumulative increase of 12.7%, 13.4% and 14.5% AP in the same order. Proposed technique has been integrated with Detectron2, MMDetection and YOLOv5 models and it is publicly available at https://github.com/obss/sahi.git .
翻译:在现场远处对小物体和物体的探测是监测应用方面的一大挑战。这些物体在图像中由少量像素代表,缺乏足够的细节,因此难以使用常规探测器进行探测。在这项工作中,提议建立一个开放源码框架,称为“剪切辅助超重推断(SAHI)”,为小物体探测提供一个通用的剪切辅助推断和微调管道。拟议的技术是通用的,因为它可以在任何现有物体探测器之上应用,而不作任何微调。实验性评估,使用Visdrone和XVView天体探测数据集的物体探测基线,表明拟议的推断方法可以分别将FCOS、VFNet和TOOD探测器的物体探测AP增加6.8%、5.1%和5.3%。此外,如果采用剪切辅助微调,检测的精确度还可以进一步提高,从而在相同的顺序中累积增加12.7%、13.4%和14.5% AP。拟议的技术已经与MMSetroviron2和YOLOV5模型相结合。