Image quality assessment (IQA) forms a natural and often straightforward undertaking for humans, yet effective automation of the task remains highly challenging. Recent metrics from the deep learning community commonly compare image pairs during training to improve upon traditional metrics such as PSNR or SSIM. However, current comparisons ignore the fact that image content affects quality assessment as comparisons only occur between images of similar content. This restricts the diversity and number of image pairs that the model is exposed to during training. In this paper, we strive to enrich these comparisons with content diversity. Firstly, we relax comparison constraints, and compare pairs of images with differing content. This increases the variety of available comparisons. Secondly, we introduce listwise comparisons to provide a holistic view to the model. By including differentiable regularizers, derived from correlation coefficients, models can better adjust predicted scores relative to one another. Evaluation on multiple benchmarks, covering a wide range of distortions and image content, shows the effectiveness of our learning scheme for training image quality assessment models.
翻译:图像质量评估(IQA)是人类的一种自然的而且往往是直接的任务,然而,任务的有效自动化仍然极具挑战性。来自深层次学习社区的最近指标通常在培训期间比较图像配对,以改进PSNR或SSIM等传统指标。然而,目前的比较忽略了这样一个事实,即图像内容影响质量评估,因为只能对类似内容的图像进行比较。这限制了模型在培训期间接触到的图像配对的多样性和数量。在本文中,我们努力丰富这些与内容多样性的比较。首先,我们放松比较限制,比较不同内容的图像配对。这增加了现有比较的种类。第二,我们采用列表比较,为模型提供一个整体的视角。通过纳入来自相关系数的不同常规,模型可以更好地调整预测的比值。对多个基准的评价,涵盖广泛的扭曲和图像内容,显示了我们培训图像质量评估模型学习计划的有效性。