Although Faster R-CNN and its variants have shown promising performance in object detection, they only exploit simple first-order representation of object proposals for final classification and regression. Recent classification methods demonstrate that the integration of high-order statistics into deep convolutional neural networks can achieve impressive improvement, but their goal is to model whole images by discarding location information so that they cannot be directly adopted to object detection. In this paper, we make an attempt to exploit high-order statistics in object detection, aiming at generating more discriminative representations for proposals to enhance the performance of detectors. To this end, we propose a novel Multi-scale Location-aware Kernel Representation (MLKP) to capture high-order statistics of deep features in proposals. Our MLKP can be efficiently computed on a modified multi-scale feature map using a low-dimensional polynomial kernel approximation.Moreover, different from existing orderless global representations based on high-order statistics, our proposed MLKP is location retentive and sensitive so that it can be flexibly adopted to object detection. Through integrating into Faster R-CNN schema, the proposed MLKP achieves very competitive performance with state-of-the-art methods, and improves Faster R-CNN by 4.9% (mAP), 4.7% (mAP) and 5.0% (AP at IOU=[0.5:0.05:0.95]) on PASCAL VOC 2007, VOC 2012 and MS COCO benchmarks, respectively. Code is available at: https://github.com/Hwang64/MLKP.
翻译:虽然快速R-CNN及其变体在物体探测方面表现良好,但它们只是利用简单第一阶的物体表示方式,为最终分类和回归提供最终分类和回归。最近的分类方法表明,将高阶统计数据纳入深层神经神经网络可以实现令人印象深刻的改进,但其目标是通过丢弃位置信息来模拟整幅图像,从而无法直接用于物体探测。在本文件中,我们试图在物体探测方面利用高阶统计数据,目的是为增强探测器性能的建议产生更具有歧视性的表述方式。为此,我们提出了一个新的多级位置-觉醒内尔代表(MLKP),以捕捉到建议中深度特征的高阶统计数据。我们的MLKP可以用低度多元内核素近光度的修改后的多尺度图进行高效的计算,从而无法直接用于物体探测。在高阶统计数据的基础上,我们提议的MLKP是位置的重复和敏感度,因此可以灵活地用于物体探测。通过在更快的 R-CN CO-aware Kern 代表(MLKP) 中,拟议的ML-RC-OMAL 和R-MAMA-% 分别在2007年的 RSAL-MA-MAP-%和2007年的S-MA-MA-MA-MA-MA-MA-%-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-SA-SA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-MA-SA-SA-SA-SA-MA-SA-SA-MA-MA-MA-MA-MA-MA-MA-SA-SA-SA-