Questing for lossy image coding (LIC) with superior efficiency on both compression performance and computation throughput is challenging. The vital factor behind is how to intelligently explore Adaptive Neighborhood Information Aggregation (ANIA) in transform and entropy coding modules. To this aim, Integrated Convolution and Self-Attention (ICSA) unit is first proposed to form content-adaptive transform to dynamically characterize and embed neighborhood information conditioned on the input. Then a Multistage Context Model (MCM) is developed to stagewisely execute context prediction using necessary neighborhood elements for accurate and parallel entropy probability estimation. Both ICSA and MCM are stacked under a Variational Auto-Encoder (VAE) architecture to derive rate-distortion optimized compact representation of input image via end-to-end training. Our method reports the superior compression performance surpassing the VVC Intra with $\approx$15% BD-rate improvement averaged across Kodak, CLIC and Tecnick datasets; and also demonstrates $\approx$10$\times$ speedup of image decoding when compared with other notable learned LIC approaches. All materials are made publicly accessible at https://njuvision.github.io/TinyLIC for reproducible research.
翻译:在压缩性能和计算完成量方面效率较高的丢失图像编码(LIC)的切换是挑战性的。关键因素是,如何在变换和加密编码模块中明智地探索适应性邻居信息聚合(ANIA) 。为此,首先建议综合变换和自我保护(ICSA)单位进行内容调整,以动态方式转换和嵌入以输入为条件的邻里信息。然后开发了一个多阶段背景模型(MCM),以便利用必要的邻里元素进行环境预测,进行准确和平行的同步概率估计。ICSA和MCM都堆叠在一个动态自动编码器自动编码(VAEE)结构下,通过端到端培训生成最佳的输入图像缩压优化缩缩缩缩缩缩缩缩图。我们的方法报告,优于 VVC Intra的压缩性表现超过了以$approx$15% BD-rate d 平均在Kodak、CLIC和Tecnick数据集中进行环境改进;还展示了美元/aprox$@LICUxxxxxxximalaldealdealalal 10 agremoduction agremoduction agnial agremodudududuductions