Nowadays, vision-based computing tasks play an important role in various real-world applications. However, many vision computing tasks, e.g. semantic segmentation, are usually computationally expensive, posing a challenge to the computing systems that are resource-constrained but require fast response speed. Therefore, it is valuable to develop accurate and real-time vision processing models that only require limited computational resources. To this end, we propose the Spatial-detail Guided Context Propagation Network (SGCPNet) for achieving real-time semantic segmentation. In SGCPNet, we propose the strategy of spatial-detail guided context propagation. It uses the spatial details of shallow layers to guide the propagation of the low-resolution global contexts, in which the lost spatial information can be effectively reconstructed. In this way, the need for maintaining high-resolution features along the network is freed, therefore largely improving the model efficiency. On the other hand, due to the effective reconstruction of spatial details, the segmentation accuracy can be still preserved. In the experiments, we validate the effectiveness and efficiency of the proposed SGCPNet model. On the Citysacpes dataset, for example, our SGCPNet achieves 69.5 % mIoU segmentation accuracy, while its speed reaches 178.5 FPS on 768x1536 images on a GeForce GTX 1080 Ti GPU card.
翻译:目前,基于视觉的计算任务在现实世界的各种应用中起着重要作用。然而,许多视觉计算任务,例如语义分割,通常计算成本昂贵,对资源受限制但需要快速反应速度的计算系统构成挑战,因此,开发精确和实时的视觉处理模型非常宝贵,只需要有限的计算资源。为此,我们提议建立空间脱尾引导背景促进网络(SGCPNet),以实现实时语义分离。在SGCPNet中,我们提出了空间脱尾引导背景传播战略。它利用浅层的空间细节来指导低分辨率全球环境的传播,在这种环境中,丢失的空间信息可以有效地重建。因此,在网络上保持高分辨率特性的必要性已经解放了,因此在很大程度上提高了模型效率。另一方面,由于空间细节的有效重建,分解准确性仍可以保持。在实验中,我们验证了拟议的SGCP网络模型模型的效能和效率。在Cellasacasacasion SIAVIC SIM5中, 其精确度为 SAFSER5 AS AS AS ASAL AS SAIC AS AS SAI SAIC SALATION SAL SAIC SAIC SAIC SAIC SAL SAIC SAIC SAL SAIC SAIC SAIC SAIC SAIC SAIC SAIC SAIC SAIC SAIC SALS SATI SATI SATI 5 SATI AL 。例例,我们 SATI SATI 。