Ultra-high resolution image segmentation has raised increasing interests in recent years due to its realistic applications. In this paper, we innovate the widely used high-resolution image segmentation pipeline, in which an ultra-high resolution image is partitioned into regular patches for local segmentation and then the local results are merged into a high-resolution semantic mask. In particular, we introduce a novel locality-aware contextual correlation based segmentation model to process local patches, where the relevance between local patch and its various contexts are jointly and complementarily utilized to handle the semantic regions with large variations. Additionally, we present a contextual semantics refinement network that associates the local segmentation result with its contextual semantics, and thus is endowed with the ability of reducing boundary artifacts and refining mask contours during the generation of final high-resolution mask. Furthermore, in comprehensive experiments, we demonstrate that our model outperforms other state-of-the-art methods in public benchmarks. Our released codes are available at https://github.com/liqiokkk/FCtL.
翻译:近年来,超高分辨率图像分解由于其现实应用,引起了越来越多的兴趣。在本文中,我们创新了广泛使用的高分辨率图像分解管道,将超高分辨率图像分解成局部分解的常规补丁,然后将当地结果合并成高分辨率语义遮罩。特别是,我们引入了一种基于地貌认知的背景相关分解新颖模式,用于处理本地补丁,当地补丁及其各种背景的相关性被联合和补充用于处理具有巨大变异的语义区域。此外,我们推出了一个背景语义精细化网络,将本地分解结果与其背景语义连接起来,从而赋予了在最终生成高分辨率面具期间减少边界文物和完善蒙面等离子的能力。此外,在全面实验中,我们证明我们的模型超越了公共基准中的其他州-艺术方法。我们发布的代码可以在 https://github.com/liqikk/FCTL上查阅。