We consider the problem of obtaining image quality representations in a self-supervised manner. We use prediction of distortion type and degree as an auxiliary task to learn features from an unlabeled image dataset containing a mixture of synthetic and realistic distortions. We then train a deep Convolutional Neural Network (CNN) using a contrastive pairwise objective to solve the auxiliary problem. We refer to the proposed training framework and resulting deep IQA model as the CONTRastive Image QUality Evaluator (CONTRIQUE). During evaluation, the CNN weights are frozen and a linear regressor maps the learned representations to quality scores in a No-Reference (NR) setting. We show through extensive experiments that CONTRIQUE achieves competitive performance when compared to state-of-the-art NR image quality models, even without any additional fine-tuning of the CNN backbone. The learned representations are highly robust and generalize well across images afflicted by either synthetic or authentic distortions. Our results suggest that powerful quality representations with perceptual relevance can be obtained without requiring large labeled subjective image quality datasets. The implementations used in this paper are available at \url{https://github.com/pavancm/CONTRIQUE}.
翻译:我们以自我监督的方式考虑获得图像质量表述的问题。我们用扭曲类型和度的预测作为辅助任务,从含有合成和现实扭曲混合的未贴标签图像数据集中学习特征,然后用对比式双向目标培训深革命神经网络(CNN)以解决辅助问题。我们把拟议的培训框架和由此产生的深度IQA模型称为监控图像质量评估仪(CONTRIQUE ) 。在评估期间,CNN重量被冻结,并用线性反射器将所学的演示图示绘制成无引用(NR)环境中的质量评分。我们通过广泛的实验显示,CONTRIQUE在与最新NR图像质量模型相比,即使不对CNN骨干作任何额外的微调,也取得了竞争性的绩效。我们所了解的表述非常有力,而且广泛涵盖了受合成或真实扭曲影响的各种图像。我们的结果表明,在不要求大标签的主观图像质量数据集的情况下,可以获取具有概念相关性的强大质量表述。我们通过广泛的实验表明,CTRIQ 所使用的实施方式是在CRI/URQ。