In recent years, deep neural network (DNN) compression systems have proved to be highly effective for designing source codes for many natural sources. However, like many other machine learning systems, these compressors suffer from vulnerabilities to distribution shifts as well as out-of-distribution (OOD) data, which reduces their real-world applications. In this paper, we initiate the study of OOD robust compression. Considering robustness to two types of ambiguity sets (Wasserstein balls and group shifts), we propose algorithmic and architectural frameworks built on two principled methods: one that trains DNN compressors using distributionally-robust optimization (DRO), and the other which uses a structured latent code. Our results demonstrate that both methods enforce robustness compared to a standard DNN compressor, and that using a structured code can be superior to the DRO compressor. We observe tradeoffs between robustness and distortion and corroborate these findings theoretically for a specific class of sources.
翻译:近年来,深神经网络压缩系统(DNN)被证明对设计许多自然源源源源源代码非常有效。 但是,与其他许多机器学习系统一样,这些压缩器在分布变化和分配外数据方面的脆弱性,这减少了它们真实世界的应用。在本论文中,我们启动了OOD强力压缩的研究。考虑到两种模棱两可(Wasserstein球和群体转换)的稳健性,我们提出了基于两种原则性方法的算法和建筑框架:一种是使用分布式机器人优化(DRO)来培训DN压缩机,另一种是使用结构化潜伏代码。我们的结果表明,两种方法都比标准的DNNN压缩机压缩机强,而使用结构化代码可以优于DRO压缩机。我们观察强性和扭曲之间的权衡,并从理论上为特定来源证实了这些结论。