Object detection from Unmanned Aerial Vehicles (UAVs) is of great importance in many aerial vision-based applications. Despite the great success of generic object detection methods, a significant performance drop is observed when applied to images captured by UAVs. This is due to large variations in imaging conditions, such as varying altitudes, dynamically changing viewing angles, and different capture times. These variations lead to domain imbalances and, thus, trained models suffering from domain bias. We demonstrate that domain knowledge is a valuable source of information and thus propose domain-aware object detectors by using freely accessible sensor data. By splitting the model into cross-domain and domain-specific parts, substantial performance improvements are achieved on multiple data sets across various models and metrics without changing the architecture. In particular, we achieve a new state-of-the-art performance on UAVDT for embedded real-time detectors. Furthermore, we create a new airborne image data set by annotating 13,713 objects in 2,900 images featuring precise altitude and viewing angle annotations.
翻译:尽管通用物体探测方法取得了巨大成功,但在应用无人驾驶飞行器所摄图像时观察到性能显著下降,这是因为成像条件差异很大,如高度不同、视野动态变化和捕捉时间不同。这些变化导致区域失衡,因此,经过培训的模型存在领域偏差。我们证明,域知识是宝贵的信息来源,因此通过使用可自由获取的传感器数据,建议使用域感应器探测器。通过将模型分为跨域和特定域部分,在各种模型和尺度的多个数据集上取得了显著的性能改进,而不改变结构。特别是,我们在UAVDT上实现了新的最新水平性能,用于嵌入实时探测器。此外,我们通过在2 900个图像中注明13 713个物体,显示精确高度和角度说明,从而建立了新的空中图像数据。