In this paper we present an end-to-end meta-learned system for image compression. Traditional machine learning based approaches to image compression train one or more neural network for generalization performance. However, at inference time, the encoder or the latent tensor output by the encoder can be optimized for each test image. This optimization can be regarded as a form of adaptation or benevolent overfitting to the input content. In order to reduce the gap between training and inference conditions, we propose a new training paradigm for learned image compression, which is based on meta-learning. In a first phase, the neural networks are trained normally. In a second phase, the Model-Agnostic Meta-learning approach is adapted to the specific case of image compression, where the inner-loop performs latent tensor overfitting, and the outer loop updates both encoder and decoder neural networks based on the overfitting performance. Furthermore, after meta-learning, we propose to overfit and cluster the bias terms of the decoder on training image patches, so that at inference time the optimal content-specific bias terms can be selected at encoder-side. Finally, we propose a new probability model for lossless compression, which combines concepts from both multi-scale and super-resolution probability model approaches. We show the benefits of all our proposed ideas via carefully designed experiments.
翻译:在本文中,我们展示了一个图像压缩端到端的元学系统; 传统的机器学习方法对图像压缩列车一个或多个神经网络进行一般化性能的压缩。 但是, 在推论时间, 编码器的编码器或潜伏电动输出可以优化为每个测试图像。 这种优化可以被视为一种适应或比输入内容更适应的形式。 为了缩小培训与推断条件之间的差距, 我们提议了一种新的学习图像压缩培训模式, 其基础是元化学习。 在第一阶段, 神经网络通常受到培训。 在第二阶段, 模型- 高级元化学习方法可以适应图像压缩的具体案例, 使内部环显示潜伏的变色器超配, 外环可以更新基于过度调整性能的编码器和解析器神经网络。 此外, 在进行元学习后, 我们建议过度调整和集中该解析器的偏差术语, 从而在推断时间里, 神经网络网络网络网络网络的神经网络网络网络通常会被训练整齐。 最佳内容- 元化的元化模型最终可以选择通过超标的概率 。