It was estimated that the world produced $59 ZB$ ($5.9 \times 10^{13} GB$) of data in 2020, resulting in the enormous costs of both data storage and transmission. Fortunately, recent advances in deep generative models have spearheaded a new class of so-called "neural compression" algorithms, which significantly outperform traditional codecs in terms of compression ratio. Unfortunately, the application of neural compression garners little commercial interest due to its limited bandwidth; therefore, developing highly efficient frameworks is of critical practical importance. In this paper, we discuss lossless compression using normalizing flows which have demonstrated a great capacity for achieving high compression ratios. As such, we introduce iFlow, a new method for achieving efficient lossless compression. We first propose Modular Scale Transform (MST) and a novel family of numerically invertible flow transformations based on MST. Then we introduce the Uniform Base Conversion System (UBCS), a fast uniform-distribution codec incorporated into iFlow, enabling efficient compression. iFlow achieves state-of-the-art compression ratios and is $5\times$ quicker than other high-performance schemes. Furthermore, the techniques presented in this paper can be used to accelerate coding time for a broad class of flow-based algorithms.
翻译:据估算,2020年全世界生产了59 ZB$(5.9美元乘以10 ⁇ 13美元)的数据,这导致数据存储和传输费用高昂。幸运的是,深基因模型最近的进展引出了一种新的所谓的“神经压缩”算法,在压缩比例方面大大优于传统代码。不幸的是,神经压缩的应用因其带宽有限而没有多少商业利益;因此,开发高效的框架具有至关重要的实际重要性。在本文中,我们讨论使用正常化流进行无损压缩的问题,这些流显示出实现高压缩比率的巨大能力。因此,我们引入iFlow,这是实现高效无损压缩的新方法。我们首先提议以MST为基础的“神经压缩”算法转型(MST)和在数字上不可忽略的流变新组合。然后我们引入统一基转换系统(UBCS),这是一套快速统一分配的代码,被纳入iFlow,能够实现高效压缩。iFlow实现了状态的压缩比率,并且比其他高压压速度的高级算法计划更快。此外,在高压压法中,这个技术可以比其他高速度更快。