Most existing CNN-based super-resolution (SR) methods are developed based on an assumption that the degradation is fixed and known (e.g., bicubic downsampling). However, these methods suffer a severe performance drop when the real degradation is different from their assumption. To handle various unknown degradations in real-world applications, previous methods rely on degradation estimation to reconstruct the SR image. Nevertheless, degradation estimation methods are usually time-consuming and may lead to SR failure due to large estimation errors. In this paper, we propose an unsupervised degradation representation learning scheme for blind SR without explicit degradation estimation. Specifically, we learn abstract representations to distinguish various degradations in the representation space rather than explicit estimation in the pixel space. Moreover, we introduce a Degradation-Aware SR (DASR) network with flexible adaption to various degradations based on the learned representations. It is demonstrated that our degradation representation learning scheme can extract discriminative representations to obtain accurate degradation information. Experiments on both synthetic and real images show that our network achieves state-of-the-art performance for the blind SR task. Code is available at: https://github.com/LongguangWang/DASR.
翻译:现有大多数基于CNN的超分辨率(SR)方法的制定所依据的假设是,降解是固定的和已知的(例如,双立下取样),然而,当实际降解不同于其假设时,这些方法的性能严重下降。为了处理现实应用中各种未知的退化,以前的方法依靠退化估计来重建SR形象。然而,退化估计方法通常耗费时间,并可能导致斯洛伐克共和国的退化估计错误。在本文中,我们建议为盲人SR制定一个不受监督的退化代表学习计划,但没有明确的降解估计。具体地说,我们学习抽象的表述,以区分代表空间中的各种退化,而不是像素空间中的明确估计。此外,我们引入了一个退化-Awarard SR(DASR)网络,根据所学的表述灵活地适应各种退化。我们退化代表学习计划可以提取歧视的表述,以获得准确的降解信息。对合成图像和真实图像的实验表明,我们的网络在盲SR任务中实现了状态-艺术性表现。代码见: https://giuth/Wang/DADA/DADA/DADA/CO。