A fundamental question in designing lossy data compression schemes is how well one can do in comparison with the rate-distortion function, which describes the known theoretical limits of lossy compression. Motivated by the empirical success of deep neural network (DNN) compressors on large, real-world data, we investigate methods to estimate the rate-distortion function on such data, which would allow comparison of DNN compressors with optimality. While one could use the empirical distribution of the data and apply the Blahut-Arimoto algorithm, this approach presents several computational challenges and inaccuracies when the datasets are large and high-dimensional, such as the case of modern image datasets. Instead, we re-formulate the rate-distortion objective, and solve the resulting functional optimization problem using neural networks. We apply the resulting rate-distortion estimator, called NERD, on popular image datasets, and provide evidence that NERD can accurately estimate the rate-distortion function. Using our estimate, we show that the rate-distortion achievable by DNN compressors are within several bits of the rate-distortion function for real-world datasets. Additionally, NERD provides access to the rate-distortion achieving channel, as well as samples from its output marginal. Therefore, using recent results in reverse channel coding, we describe how NERD can be used to construct an operational one-shot lossy compression scheme with guarantees on the achievable rate and distortion. Experimental results demonstrate competitive performance with DNN compressors.
翻译:在设计损失数据压缩计划时,一个根本的问题是,与描述已知损失压缩理论极限的速率扭曲功能相比,人们能够做得如何。受深神经网络压缩机在大型真实世界数据上的经验性成功驱动,我们调查了在这类数据上估计率扭曲功能的方法,这样可以将DNN压缩机与最佳性能进行比较。虽然可以使用数据的经验分布,并应用Blahut-Arimoto算法,但这一方法在数据集大而高的高度(例如现代图像数据集的例子)时,提出了若干计算挑战和不准确性。相反,我们重新构建了率扭曲目标,并用神经网络解决由此产生的功能优化问题。我们用流行图像数据集应用了标准扭曲器,并提供了证据,证明NERD可以准确估算汇率扭曲功能。我们用我们估算的估算, 将Rest- droal- developal 的最近率扭曲率与DNNNC-ral-deal-deal-ral-ral-ral-ral-ral-ral-ral-ral-ral-ral-ral-ral-de-ral-ral-ral-ral-ral-ral-ral-ral-ral-de-ral-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-lation-s-l-l-l-l-l-lation-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l-l