Despite the growing popularity of video super-resolution (VSR), there is still no good way to assess the quality of the restored details in upscaled frames. Some SR methods may produce the wrong digit or an entirely different face. Whether a method's results are trustworthy depends on how well it restores truthful details. Image super-resolution can use natural distributions to produce a high-resolution image that is only somewhat similar to the real one. VSR enables exploration of additional information in neighboring frames to restore details from the original scene. The ERQA metric, which we propose in this paper, aims to estimate a model's ability to restore real details using VSR. On the assumption that edges are significant for detail and character recognition, we chose edge fidelity as the foundation for this metric. Experimental validation of our work is based on the MSU Video Super-Resolution Benchmark, which includes the most difficult patterns for detail restoration and verifies the fidelity of details from the original frame. Code for the proposed metric is publicly available at https://github.com/msu-video-group/ERQA.
翻译:尽管视频超分辨率(VSR)越来越受欢迎,但仍然无法很好地评估升级框中恢复细节的质量。有些SR方法可能会产生错误的数字或完全不同的面孔。一种方法的结果是否值得信赖取决于它恢复真实细节的程度。图像超分辨率可以使用自然分布来制作高分辨率图像,而这种图像只与真实图像略相近。VSR能够探索邻接框中的额外信息,以恢复原始场景的细节。我们在本文件中提议的ERQA指标旨在估计模型使用VSR恢复真实细节的能力。假设边缘对细节和特征识别很重要,我们选择边缘忠诚作为这一指标的基础。我们工作的实验性验证以MSU视频超分辨率基准为基础,其中包括详细恢复和核实原始框架详细细节的最困难模式。拟议指标的代码可在https://github.com/msu-dVicio-group/ERQA上公开查阅。