Learning to detect objects, such as humans, in imagery captured by an unmanned aerial vehicle (UAV) usually suffers from tremendous variations caused by the UAV's position towards the objects. In addition, existing UAV-based benchmark datasets do not provide adequate dataset metadata, which is essential for precise model diagnosis and learning features invariant to those variations. In this paper, we introduce Archangel, the first UAV-based object detection dataset composed of real and synthetic subsets captured with similar imagining conditions and UAV position and object pose metadata. A series of experiments are carefully designed with a state-of-the-art object detector to demonstrate the benefits of leveraging the metadata during model evaluation. Moreover, several crucial insights involving both real and synthetic data during model fine-tuning are presented. In the end, we discuss the advantages, limitations, and future directions regarding Archangel to highlight its distinct value for the broader machine learning community.
翻译:在无人驾驶航空器(无人驾驶航空器)所摄图像中,学习探测物体,例如人类,通常会因无人驾驶航空器对天体的定位而发生巨大变化;此外,现有以无人驾驶航空器为基础的基准数据集没有提供足够的数据集元数据,而这些数据对于精确的模型诊断和学习特征是不可或缺的,而对于这些变异也是不可或缺的。在本文中,我们介绍第一组以无人驾驶航空器为基础的天体探测数据集Archangel,这是由以类似想象条件所捕捉的真子和合成子集组成的首个天体探测数据集,无人驾驶航空器的位置和天体构成元数据。经过精心设计的一系列实验用最先进的天体探测器来展示在模型评估期间利用元数据的好处。此外,还介绍了一些涉及模型调整期间真实和合成数据的重要见解。最后,我们讨论了天使的优势、局限性和未来方向,以突出其对更广泛的机器学习界的独特价值。