Accurately ranking the vast number of candidate detections is crucial for dense object detectors to achieve high performance. Prior work uses the classification score or a combination of classification and predicted localization scores to rank candidates. However, neither option results in a reliable ranking, thus degrading detection performance. In this paper, we propose to learn an Iou-aware Classification Score (IACS) as a joint representation of object presence confidence and localization accuracy. We show that dense object detectors can achieve a more accurate ranking of candidate detections based on the IACS. We design a new loss function, named Varifocal Loss, to train a dense object detector to predict the IACS, and propose a new star-shaped bounding box feature representation for IACS prediction and bounding box refinement. Combining these two new components and a bounding box refinement branch, we build an IoU-aware dense object detector based on the FCOS+ATSS architecture, that we call VarifocalNet or VFNet for short. Extensive experiments on MS COCO show that our VFNet consistently surpasses the strong baseline by $\sim$2.0 AP with different backbones. Our best model VFNet-X-1200 with Res2Net-101-DCN achieves a single-model single-scale AP of 55.1 on COCO test-dev, which is state-of-the-art among various object detectors.Code is available at https://github.com/hyz-xmaster/VarifocalNet .
翻译:精确地排列大量候选人检测数据对于密集物体探测器取得高性能至关重要。 先前的工作使用分类分或分类和预测本地化分数的组合来对候选人进行排名。 但是, 两种选项都没有产生可靠的排名, 从而降低检测性能 。 在本文中, 我们提议学习一个岩- 水分类分分分数(IACS), 以联合表示物体存在信心和本地化准确性能。 我们显示, 密度物体探测器能够根据IACS 来对候选人检测进行更准确的排序。 我们设计了一个新的损失函数, 名为Varifocal Loss, 以训练一个密度物体探测器来预测IACS, 并为IACS 预测和捆绑箱的改进提出一个新的恒星型捆绑盒特征表示。 将这两个新组件和捆绑式箱改进分支结合起来, 我们根据FCOS+ATSS的架构, 建立一个IOU- 甚浓度物体探测器, 我们叫VarifalNet 或VFNet 。 MSCO 的大规模实验显示, 我们的VFNet在VN- ISO- cal 基线中持续超过由$10- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- ASl ASl AS- AS- ASU_ AS- AS- AS- AS- AS- AS- AS_ AS- AS- ASl_ ASl AS- ASU_ AS_ AS- ASl_ ASl AS AS AS AS_ AS_ AS_ AS_ AS_ AS_ AS_ ASU_ AS_ ASU_ ASU_ AS_ AS_ AS_ AS_ AS_ AS_ ASl IS- ASl ASl ASl ASl ASl AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS- AS