Generative models for image restoration, enhancement, and generation have significantly improved the quality of the generated images. Surprisingly, these models produce more pleasant images to the human eye than other methods, yet, they may get a lower perceptual quality score using traditional perceptual quality metrics such as PSNR or SSIM. Therefore, it is necessary to develop a quantitative metric to reflect the performance of new algorithms, which should be well-aligned with the person's mean opinion score (MOS). Learning-based approaches for perceptual image quality assessment (IQA) usually require both the distorted and reference image for measuring the perceptual quality accurately. However, commonly only the distorted or generated image is available. In this work, we explore the performance of transformer-based full-reference IQA models. We also propose a method for IQA based on semi-supervised knowledge distillation from full-reference teacher models into blind student models using noisy pseudo-labeled data. Our approaches achieved competitive results on the NTIRE 2022 Perceptual Image Quality Assessment Challenge: our full-reference model was ranked 4th, and our blind noisy student was ranked 3rd among 70 participants, each in their respective track.
翻译:图像恢复、增强和生成的生成模型显著提高了图像生成的质量。 令人惊讶的是,这些模型产生的人类眼部图像比其他方法更令人愉快,然而,它们可能利用PSNR 或 SSIM 等传统感知质量衡量标准获得较低的感官质量评分。 因此,有必要开发一个定量衡量标准,以反映新算法的性能,这种算法应当与个人的平均观点评分(MOS)相一致。 视觉图像质量评估(IQA)的学习方法通常需要扭曲的和参考的图像来准确测量感官质量。 然而,通常只有扭曲或生成的图像才能得到。 在这项工作中,我们探索基于变异器的全参考 IQA 模型的性能。 我们还提议了一个基于半受监督的知识提炼的IQA方法,从全参考教师模型进入盲人学生模型,使用杂音假标签数据。 我们在2022 NTIRE 视觉图像质量评估(IQA) 中取得了竞争性结果。 我们的全参考模型在每一个盲人学生中排名第70位中位。