Generative Adversarial Networks (GANs) have achieved huge success in generating high-fidelity images, however, they suffer from low efficiency due to tremendous computational cost and bulky memory usage. Recent efforts on compression GANs show noticeable progress in obtaining smaller generators by sacrificing image quality or involving a time-consuming searching process. In this work, we aim to address these issues by introducing a teacher network that provides a search space in which efficient network architectures can be found, in addition to performing knowledge distillation. First, we revisit the search space of generative models, introducing an inception-based residual block into generators. Second, to achieve target computation cost, we propose a one-step pruning algorithm that searches a student architecture from the teacher model and substantially reduces searching cost. It requires no l1 sparsity regularization and its associated hyper-parameters, simplifying the training procedure. Finally, we propose to distill knowledge through maximizing feature similarity between teacher and student via an index named Global Kernel Alignment (GKA). Our compressed networks achieve similar or even better image fidelity (FID, mIoU) than the original models with much-reduced computational cost, e.g., MACs. Code will be released at https://github.com/snap-research/CAT.
翻译:在这项工作中,我们的目标是通过引进教师网络,提供搜索空间,除了进行知识蒸馏外,还能找到高效的网络结构,从而解决这些问题。首先,我们重新审视基因模型的搜索空间,在发电机中引入基于初始的残余块。第二,为了实现目标计算成本,我们建议采用一步的运行算法,从教师模型中搜索学生结构,并大幅降低搜索成本。这要求不要求I1缩放规范及其相关的超参数,简化培训程序。最后,我们提议通过名为Global Kernel Conness(GKA)的指数,通过最大限度地提高师生之间的特征相似性来挖掘知识。我们压缩的网络实现了类似甚至更好的图像真实性(FID, mIOU),比原始的MAC-codeal 模型(MACMAC/CAT)要多得多的成本计算。