Recent deep learning methods have achieved promising results in image shadow removal. However, most of the existing approaches focus on working locally within shadow and non-shadow regions, resulting in severe artifacts around the shadow boundaries as well as inconsistent illumination between shadow and non-shadow regions. It is still challenging for the deep shadow removal model to exploit the global contextual correlation between shadow and non-shadow regions. In this work, we first propose a Retinex-based shadow model, from which we derive a novel transformer-based network, dubbed ShandowFormer, to exploit non-shadow regions to help shadow region restoration. A multi-scale channel attention framework is employed to hierarchically capture the global information. Based on that, we propose a Shadow-Interaction Module (SIM) with Shadow-Interaction Attention (SIA) in the bottleneck stage to effectively model the context correlation between shadow and non-shadow regions. We conduct extensive experiments on three popular public datasets, including ISTD, ISTD+, and SRD, to evaluate the proposed method. Our method achieves state-of-the-art performance by using up to 150X fewer model parameters.
翻译:最近的深层学习方法在清除图像阴影方面取得了可喜的成果;然而,大多数现有方法侧重于在阴影和非阴影区域内在当地开展工作,导致阴影边界周围有严重的文物,阴影和非阴影区域之间有不一致的照明;对于深阴影清除模型来说,利用阴影和非阴影区域之间的全球背景关系,仍然具有挑战性;在这项工作中,我们首先提议基于Retinex的影子模型,从中我们产生一个新的基于变压器的网络,称为Shandow Former,以利用非阴影区域来帮助恢复阴影区域;采用多层次的频道关注框架来从等级上捕捉全球信息;在此基础上,我们提议在瓶颈阶段采用一个带有阴影与非阴影区域间注意的影子互动模块(SIM),以有效地模拟阴影与非阴影区域之间的背景关系;我们在三个广受欢迎的公共数据集上进行广泛的实验,包括ISTD、ISTD+和SRD,以评价拟议的方法。我们的方法通过使用多达150x的参数实现状态。