Learned Compression (LC) is the emerging technology for compressing image and video content, using deep neural networks. Despite being new, LC methods have already gained a compression efficiency comparable to state-of-the-art image compression, such as HEVC or even VVC. However, the existing solutions often require a huge computational complexity, which discourages their adoption in international standards or products. This paper provides a comprehensive complexity assessment of several notable methods, that shed light on the matter, and guide the future development of this field by presenting key findings. To do so, six existing methods have been evaluated for both encoding and decoding, on CPU and GPU platforms. Various aspects of complexity such as the overall complexity, share of each coding module, number of operations, number of parameters, most demanding GPU kernels, and memory requirements have been measured and compared on Kodak dataset. The reported results (1) quantify the complexity of LC methods, (2) fairly compare different methods, and (3) a major contribution of the work is identifying and quantifying the key factors affecting the complexity.
翻译:利用深神经网络压缩图像和视频内容的新兴技术(LC)是利用深神经网络压缩图像和视频内容的新兴技术。尽管LC方法是新的,但已经取得了与最新图像压缩(如HEVC甚至VVC)相当的压缩效率。然而,现有的解决方案往往需要巨大的计算复杂性,这不利于在国际标准或产品中采用这些技术。本文对若干值得注意的方法进行了全面复杂的评估,这些方法揭示了问题,并通过提出关键结论来指导该领域的未来发展。为此,对CPU和GPU平台的编码和解码工作,已经对六种现有方法进行了评估。复杂性的各个方面,如总体复杂性、每个编码模块的份额、操作数量、参数数量、最严格的GPU内核和记忆要求,已经对Kodak数据集进行了衡量和比较。所报告的结果(1) 量化LC方法的复杂性,(2) 比较不同的方法,以及(3) 这项工作的主要贡献是确定和量化影响复杂性的关键因素。