Within the field of instance segmentation, most of the state-of-the-art deep learning networks rely nowadays on cascade architectures, where multiple object detectors are trained sequentially, re-sampling the ground truth at each step. This offers a solution to the problem of exponentially vanishing positive samples. However, it also translates into an increase in network complexity in terms of the number of parameters. To address this issue, we propose Recursively Refined R-CNN (R^3-CNN) which avoids duplicates by introducing a loop mechanism instead. At the same time, it achieves a quality boost using a recursive re-sampling technique, where a specific IoU quality is utilized in each recursion to eventually equally cover the positive spectrum. Our experiments highlight the specific encoding of the loop mechanism in the weights, requiring its usage at inference time. The R^3-CNN architecture is able to surpass the recently proposed HTC model, while reducing the number of parameters significantly. Experiments on COCO minival 2017 dataset show performance boost independently from the utilized baseline model. The code is available online at https://github.com/IMPLabUniPr/mmdetection/tree/r3_cnn.
翻译:在实例分割领域,目前大多数最先进的深层学习网络都依赖级联结构,多物体探测器是按顺序对级联结构进行连续培训,对每步的地面真相进行重新取样,这为解决指数性消失正数样本的问题提供了解决办法,但从参数数量来看,这也转化为网络复杂性的增加。为解决这一问题,我们提议重新精炼R-CNN(R3-CNN),通过采用环路机制避免重叠。与此同时,它利用循环性再取样技术实现质量提升,每次循环性重新取样都使用特定的IoU质量,最终同样覆盖正谱。我们的实验突出了重量圈机制的具体编码,需要从推移时使用。R3-CNN结构能够超过最近提议的HTC模型,同时显著减少参数数量。CO微型2017数据库实验显示业绩提升,与使用的基准模型无关。该代码可在线查阅 httpssm/trealrum/treal-mbrum/Umbrealmm/Angrm_nnth。