Structured reconstruction is a non-trivial dense prediction problem, which extracts structural information (\eg, building corners and edges) from a raster image, then reconstructs it to a 2D planar graph accordingly. Compared with common segmentation or detection problems, it significantly relays on the capability that leveraging holistic geometric information for structural reasoning. Current transformer-based approaches tackle this challenging problem in a two-stage manner, which detect corners in the first model and classify the proposed edges (corner-pairs) in the second model. However, they separate two-stage into different models and only share the backbone encoder. Unlike the existing modeling strategies, we present an enhanced corner representation method: 1) It fuses knowledge between the corner detection and edge prediction by sharing feature in different granularity; 2) Corner candidates are proposed in four heatmap channels w.r.t its direction. Both qualitative and quantitative evaluations demonstrate that our proposed method can better reconstruct fine-grained structures, such as adjacent corners and tiny edges. Consequently, it outperforms the state-of-the-art model by +1.9\%@F-1 on Corner and +3.0\%@F-1 on Edge.
翻译:结构化重建是一个非常棘手的密集预测问题,它从光栅图像中提取结构信息(例如建筑角和边缘),然后相应地重建为2D平面图。与常见的分割或检测问题相比,它在于依赖于利用整体几何信息进行结构推理的能力。当前基于Transformer的方法通过两阶段的方式解决这个具有挑战性的问题,第一阶段检测角点,第二阶段分类提出的边缘(角点对)。然而,它们将两个阶段分开并仅共享骨干编码器。与现有的建模策略不同,我们提出了增强的角点表示方法:1)通过在不同粒度上共享特征,融合角点检测和边缘预测之间的知识;2)与其方向相对应,在四个热图通道中提出角点候选。定量和定性评估都证明,我们提出的方法能够更好地重建细粒度结构,例如相邻的角点和微小的边缘。因此,在Corner和Edge上,它的F-1值分别比最先进的模型高出+1.9\%和+3.0\%。