Assuming a known degradation model, the performance of a learned image super-resolution (SR) model depends on how well the variety of image characteristics within the training set matches those in the test set. As a result, the performance of an SR model varies noticeably from image to image over a test set depending on whether characteristics of specific images are similar to those in the training set or not. Hence, in general, a single SR model cannot generalize well enough for all types of image content. In this work, we show that training multiple SR models for different classes of images (e.g., for text, texture, etc.) to exploit class-specific image priors and employing a post-processing network that learns how to best fuse the outputs produced by these multiple SR models surpasses the performance of state-of-the-art generic SR models. Experimental results clearly demonstrate that the proposed multiple-model SR (MMSR) approach significantly outperforms a single pre-trained state-of-the-art SR model both quantitatively and visually. It even exceeds the performance of the best single class-specific SR model trained on similar text or texture images.
翻译:假设已知的降解模型,所学的图像超分辨率(SR)模型的性能取决于培训组内图像特征的多样性与测试组内图像特征的差别,因此,根据具体图像的特性是否与培训组内的图像相类似,在测试组内,SR模型的性能在图像和图像上明显不同,因此,一般来说,单一的SR模型无法充分概括所有类型的图像内容。在这项工作中,我们表明,为不同类别图像(如文字、纹理等)培训多个SR模型,以利用特定类图像的前期,并使用后处理网络,学习如何最好地结合这些多类SR模型产生的产出,超过最先进的通用SR模型的性能。实验结果清楚地表明,拟议的多类SR模型方法在数量和视觉上都大大超越了单一的经过培训的、最特定类的SR模型的性能。