User data confidentiality protection is becoming a rising challenge in the present deep learning research. Without access to data, conventional data-driven model compression faces a higher risk of performance degradation. Recently, some works propose to generate images from a specific pretrained model to serve as training data. However, the inversion process only utilizes biased feature statistics stored in one model and is from low-dimension to high-dimension. As a consequence, it inevitably encounters the difficulties of generalizability and inexact inversion, which leads to unsatisfactory performance. To address these problems, we propose MixMix based on two simple yet effective techniques: (1) Feature Mixing: utilizes various models to construct a universal feature space for generalized inversion; (2) Data Mixing: mixes the synthesized images and labels to generate exact label information. We prove the effectiveness of MixMix from both theoretical and empirical perspectives. Extensive experiments show that MixMix outperforms existing methods on the mainstream compression tasks, including quantization, knowledge distillation, and pruning. Specifically, MixMix achieves up to 4% and 20% accuracy uplift on quantization and pruning, respectively, compared to existing data-free compression work.
翻译:在目前的深层学习研究中,用户数据保密保护正在成为一项日益严峻的挑战。没有数据,常规数据驱动模型压缩将面临更高的性能退化风险。最近,有些作品提议从特定预先培训的模型中生成图像,作为培训数据。然而,自转过程只使用一个模型中储存的偏差特征统计数据,从低差异到高差异。结果,它不可避免地遇到一般性和不实际的转换的困难,导致不令人满意的性能。为了解决这些问题,我们提议混合混合基于两种简单而有效的技术:(1) 特性混合:利用各种模型构建通用的反转通用特征空间;(2) 数据混合:将合成图像和标签混在一起,以生成准确的标签信息。我们从理论和经验角度证明混合的有效性。广泛的实验显示,MixMixMix超越了主流压缩任务的现有方法,包括裁剪、知识蒸馏和重新运行。具体地说,MixMix将分别实现4%和20%的不固定化,将现有数据进行对比,将压缩到平整。