As multimedia content often contains noise from intrinsic defects of digital devices, image denoising is an important step for high-level vision recognition tasks. Although several studies have developed the denoising field employing advanced Transformers, these networks are too momory-intensive for real-world applications. Additionally, there is a lack of research on lightweight denosing (LWDN) with Transformers. To handle this, this work provides seven comparative baseline Transformers for LWDN, serving as a foundation for future research. We also demonstrate the parts of randomly cropped patches significantly affect the denoising performances during training. While previous studies have overlooked this aspect, we aim to train our baseline Transformers in a truly fair manner. Furthermore, we conduct empirical analyses of various components to determine the key considerations for constructing LWDN Transformers. Codes are available at https://github.com/rami0205/LWDN.
翻译:随着多媒体内容通常包含数字设备固有缺陷导致的噪声,图像去噪对于高级视觉识别任务非常重要。尽管一些研究已经使用先进的Transformer开发了去噪领域,但这些网络对于实际应用来说太过内存密集。此外,缺少对采用Transformer的轻量级去噪(LWDN)的研究。为了解决这个问题,本文提供了7个适用于LWDN的比较基准Transformer,为未来的研究奠定了基础。我们还演示了随机裁剪图像补丁的部分显著影响训练中的去噪性能。尽管以前的研究忽略了这个方面,我们旨在以真正公正的方式训练基准Transformer。此外,我们进行了多种组件的实证分析,以确定构建LWDN Transformer的关键考虑因素。代码可在https://github.com/rami0205/LWDN获取。