The choice of a loss function is an important factor when training neural networks for image restoration problems, such as single image super resolution. The loss function should encourage natural and perceptually pleasing results. A popular choice for a loss is a pre-trained network, such as VGG, which is used as a feature extractor for computing the difference between restored and reference images. However, such an approach has multiple drawbacks: it is computationally expensive, requires regularization and hyper-parameter tuning, and involves a large network trained on an unrelated task. Furthermore, it has been observed that there is no single loss function that works best across all applications and across different datasets. In this work, we instead propose to train a set of loss functions that are application specific in nature. Our loss function comprises a series of discriminators that are trained to detect and penalize the presence of application-specific artifacts. We show that a single natural image and corresponding distortions are sufficient to train our feature extractor that outperforms state-of-the-art loss functions in applications like single image super resolution, denoising, and JPEG artifact removal. Finally, we conclude that an effective loss function does not have to be a good predictor of perceived image quality, but instead needs to be specialized in identifying the distortions for a given restoration method.
翻译:选择损失函数是培训神经网络解决图像恢复问题的重要因素, 如单一图像超分辨率 。 损失函数应该鼓励自然和感知上令人愉快的结果 。 对损失的流行选择是一个预先训练的网络, 如 VGG, 用作计算恢复图像和参考图像差异的特征提取器。 但是, 这种方法有多重缺点: 它计算成本昂贵, 需要正规化和超参数调, 并需要经过一个大型网络来培训一个不相关的任务。 此外, 人们发现, 在所有应用程序和不同数据集中, 没有任何单一的损失函数最有效发挥作用。 在这项工作中, 我们提议培训一组特定性质的损失函数。 我们的损失函数由一系列受过训练的歧视问题组成, 以检测和惩罚应用特定文物的存在。 我们显示, 单一的自然图像和相应的扭曲足以训练我们的特征提取器, 从而超越了不相干的任务。 此外, 已经发现, 在应用中, 单个图像超级分辨率、 解译和 JEG 文物清除中, 没有一个最先进的损失函数。 最后, 我们的结论是, 确定一个有效的修复方法是有效的质量 。