Recent single-image super-resolution (SISR) networks, which can adapt their network parameters to specific input images, have shown promising results by exploiting the information available within the input data as well as large external datasets. However, the extension of these self-supervised SISR approaches to video handling has yet to be studied. Thus, we present a new learning algorithm that allows conventional video super-resolution (VSR) networks to adapt their parameters to test video frames without using the ground-truth datasets. By utilizing many self-similar patches across space and time, we improve the performance of fully pre-trained VSR networks and produce temporally consistent video frames. Moreover, we present a test-time knowledge distillation technique that accelerates the adaptation speed with less hardware resources. In our experiments, we demonstrate that our novel learning algorithm can fine-tune state-of-the-art VSR networks and substantially elevate performance on numerous benchmark datasets.
翻译:最近的单一图像超分辨率(SISSR)网络可以将其网络参数调整为特定输入图像,通过利用输入数据以及大型外部数据集中现有的信息,显示了有希望的成果。然而,尚未研究这些自监督的SISSR方法对视频处理的延伸。因此,我们提出了一个新的学习算法,允许传统视频超分辨率(VSR)网络在不使用地面实况数据集的情况下调整其参数以测试视频框架。通过在空间和时间上使用许多自相近的间隔,我们改进了经过充分培训的VSR网络的性能,并制作了时间一致的视频框架。此外,我们展示了一种测试时知识蒸馏技术,以较少的硬件资源加速适应速度。在我们的实验中,我们证明我们的新学习算法可以微调目前先进的VSR网络的状态,并大幅提升许多基准数据集的性能。