In recent years, video compression techniques have been significantly challenged by the rapidly increased demands associated with high quality and immersive video content. Among various compression tools, post-processing can be applied on reconstructed video content to mitigate visible compression artefacts and to enhance overall perceptual quality. Inspired by advances in deep learning, we propose a new CNN-based post-processing approach, which has been integrated with two state-of-the-art coding standards, VVC and AV1. The results show consistent coding gains on all tested sequences at various spatial resolutions, with average bit rate savings of 4.0% and 5.8% against original VVC and AV1 respectively (based on the assessment of PSNR). This network has also been trained with perceptually inspired loss functions, which have further improved reconstruction quality based on perceptual quality assessment (VMAF), with average coding gains of 13.9% over VVC and 10.5% against AV1.
翻译:近年来,录相压缩技术因高质量和消化性视频内容的需求迅速增长而面临重大挑战,在各种压缩工具中,后处理可应用于重建视频内容,以减少可见压缩工艺,提高总体感官质量。在深层次学习进展的启发下,我们提议采用新的有线电视新闻网后处理方法,该方法已经与VVC和AV1两种最先进的编码标准相结合。 结果表明,各种空间分辨率的所有测试序列均取得了一致的编码收益,相对于原VVC和AV1(根据PSNR的评估),平均比VC和AV1分别节省4.0%和5.8%的位数率。这个网络还接受了感知性损失功能的培训,根据感性质量评估(VMAF)进一步提高了重建质量,平均比VVC和AV1的平均编码收益为13.9%,比AV1提高了1的10.5%。