This paper presents an iterative training of neural networks for intra prediction in a block-based image and video codec. First, the neural networks are trained on blocks arising from the codec partitioning of images, each paired with its context. Then, iteratively, blocks are collected from the partitioning of images via the codec including the neural networks trained at the previous iteration, each paired with its context, and the neural networks are retrained on the new pairs. Thanks to this training, the neural networks can learn intra prediction functions that both stand out from those already in the initial codec and boost the codec in terms of rate-distortion. Moreover, the iterative process allows the design of training data cleansings essential for the neural network training. When the iteratively trained neural networks are put into H.265 (HM-16.15), -4.2% of mean dB-rate reduction is obtained. By moving them into H.266 (VTM-5.0), the mean dB-rate reduction reaches -1.9%.
翻译:本文展示了对神经网络的迭接培训,以在成块图像和视频编码中进行内部预测。 首先,神经网络就图像的编码分隔产生的区块进行了培训,每个区块与上下文配对。 然后,通过编码,从图像的分区中收集区块,包括以前迭代中训练的神经网络,每个区块与上下文配对,神经网络在新配对上重新接受培训。通过这一培训,神经网络可以学习内部预测功能,这些功能与最初的编码中已有的功能相比,在速度扭曲方面提升代码。此外,迭接过程允许设计神经网络培训所需的培训数据清理。当迭接训练的神经网络被投入H.265(HM-16.15)时,获得4.2%的平均 dB-节率削减。通过将其移入H.266(VTM-5.0),平均的 dB节率削减达到-1.9%。