The past decade has witnessed significant progress on detecting objects in aerial images that are often distributed with large scale variations and arbitrary orientations. However most of existing methods rely on heuristically defined anchors with different scales, angles and aspect ratios and usually suffer from severe misalignment between anchor boxes and axis-aligned convolutional features, which leads to the common inconsistency between the classification score and localization accuracy. To address this issue, we propose a Single-shot Alignment Network (S$^2$A-Net) consisting of two modules: a Feature Alignment Module (FAM) and an Oriented Detection Module (ODM). The FAM can generate high-quality anchors with an Anchor Refinement Network and adaptively align the convolutional features according to the anchor boxes with a novel Alignment Convolution. The ODM first adopts active rotating filters to encode the orientation information and then produces orientation-sensitive and orientation-invariant features to alleviate the inconsistency between classification score and localization accuracy. Besides, we further explore the approach to detect objects in large-size images, which leads to a better trade-off between speed and accuracy. Extensive experiments demonstrate that our method can achieve state-of-the-art performance on two commonly used aerial objects datasets (i.e., DOTA and HRSC2016) while keeping high efficiency. The code is available at https://github.com/csuhan/s2anet.
翻译:过去十年来,在探测航空图像中的物体方面取得了显著进展,航空图像中分布的物体往往有大规模差异和任意取向,但是,大多数现有方法依赖具有不同尺度、角度和方位比率的超常定义锚,通常在锚框和轴对齐的进化特征之间出现严重不协调,导致分类分数和地方化准确性之间出现常见的不一致。为解决这一问题,我们提议建立一个单点对准网络(S$2$A-Net),由两个模块组成:一个功能调整模块(FAM)和一个面向方向的探测模块(ODM)。 FAM可以使用一个锁定点精度、角度和方位比例不同的比例、角度和方位比率,产生高质量的锚定锚定锚,并适应性地将进动特性与锁定框相匹配。ODM首先采用积极的旋转过滤器对方向信息进行编码,然后产生方向敏感和方向变异性特征,以缓解分类分数和本地化准确性之间的不一致性。此外,我们进一步探索用大型图像探测对象的方法,从而在速度和精确度之间实现更好的交易/准确性交易。