Although instance segmentation has made considerable advancement over recent years, it's still a challenge to design high accuracy algorithms with real-time performance. In this paper, we propose a real-time instance segmentation framework termed OrienMask. Upon the one-stage object detector YOLOv3, a mask head is added to predict some discriminative orientation maps, which are explicitly defined as spatial offset vectors for both foreground and background pixels. Thanks to the discrimination ability of orientation maps, masks can be recovered without the need for extra foreground segmentation. All instances that match with the same anchor size share a common orientation map. This special sharing strategy reduces the amortized memory utilization for mask predictions but without loss of mask granularity. Given the surviving box predictions after NMS, instance masks can be concurrently constructed from the corresponding orientation maps with low complexity. Owing to the concise design for mask representation and its effective integration with the anchor-based object detector, our method is qualified under real-time conditions while maintaining competitive accuracy. Experiments on COCO benchmark show that OrienMask achieves 34.8 mask AP at the speed of 42.7 fps evaluated with a single RTX 2080 Ti. The code is available at https://github.com/duwt/OrienMask.
翻译:尽管最近几年来,分化实例取得了相当大的进步,但设计具有实时性能的高精度算法仍是一项挑战。 在本文中,我们提议了一个名为 OrienMask 的实时分解框架。 在单级天体探测器YOLOv3 上,增加了一个面罩头,以预测一些具有歧视性的定向图,这些图被明确定义为前景和背景像素的空间抵消矢量。由于定向图的辨别能力,面罩可以在不需要额外前地偏移的情况下被回收。所有与同一锚大小相匹配的场标都有一个共同的方向图。这一特别共享战略降低了面罩预测的折现内存利用率,但不会丧失面罩颗粒性。根据NMS 之后的幸存框预测,可以同时从相应的方向图中以低复杂度制作。由于掩码代表的简明设计及其与基于锚的物体探测器的有效融合,我们的方法在实时条件下符合条件,同时保持竞争性的精确性。 COCO基准实验显示OrienMask 达到34.8 AS AS AS AM 2080/MAPSKSK,以42.7/ NSUD ASUT 的速, ASUAL AL AL 2080/ ASUD.