Although Convolutional Neural Networks (CNN) have made good progress in image restoration, the intrinsic equivalence and locality of convolutions still constrain further improvements in image quality. Recent vision transformer and self-attention have achieved promising results on various computer vision tasks. However, directly utilizing Transformer for image restoration is a challenging task. In this paper, we introduce an effective hybrid architecture for sand image restoration tasks, which leverages local features from CNN and long-range dependencies captured by transformer to improve the results further. We propose an efficient hybrid structure for sand dust image restoration to solve the feature inconsistency issue between Transformer and CNN. The framework complements each representation by modulating features from the CNN-based and Transformer-based branches rather than simply adding or concatenating features. Experiments demonstrate that SandFormer achieves significant performance improvements in synthetic and real dust scenes compared to previous sand image restoration methods.
翻译:虽然革命神经网络(CNN)在图像恢复方面取得了良好进展,但内在等同和变化地点仍然制约着图像质量的进一步改善。最近的视觉变异器和自我关注在各种计算机愿景任务上取得了可喜的成果。然而,直接利用变异器恢复图像是一项艰巨的任务。在本文件中,我们引入了沙图像恢复任务的有效混合结构,利用CNN和变异器捕获的长期依赖性地方特征进一步改进结果。我们提出了沙尘图像恢复高效混合结构,以解决变异器和CNN之间的特征不一致问题。该框架通过调制CNN和变异器分支的特征,而不是简单地添加或组合特征来补充每一种表现形式。实验表明,SandFormer在合成和真实的灰色场上取得了与以往的沙图像恢复方法相比的重大性能改进。</s>