Pretrained deep models hold their learnt knowledge in the form of the model parameters. These parameters act as memory for the trained models and help them generalize well on unseen data. However, in absence of training data, the utility of a trained model is merely limited to either inference or better initialization towards a target task. In this paper, we go further and extract synthetic data by leveraging the learnt model parameters. We dub them "Data Impressions", which act as proxy to the training data and can be used to realize a variety of tasks. These are useful in scenarios where only the pretrained models are available and the training data is not shared (e.g., due to privacy or sensitivity concerns). We show the applicability of data impressions in solving several computer vision tasks such as unsupervised domain adaptation, continual learning as well as knowledge distillation. We also study the adversarial robustness of the lightweight models trained via knowledge distillation using these data impressions. Further, we demonstrate the efficacy of data impressions in generating UAPs with better fooling rates. Extensive experiments performed on several benchmark datasets demonstrate competitive performance achieved using data impressions in absence of the original training data.
翻译:受过训练的深层模型以模型参数的形式保留其所学知识。这些参数作为经过训练的模型的记忆,有助于它们广泛了解无形数据。然而,在缺乏培训数据的情况下,经过训练的模型的效用仅限于推断或更好地初始化,以完成目标任务。在本文中,我们利用所学的模型参数,进一步提取合成数据。我们用它们来代替培训数据,并可用于完成各种任务。这些参数在只有经过训练的模型而且培训数据不共享(例如,由于隐私或敏感性问题)的情景中有用。我们展示了数据印象在解决若干计算机愿景任务时的适用性,例如未经监督的域适应、持续学习以及知识蒸馏等。我们还研究了通过利用这些数据的印象进行知识蒸馏所培训的轻量模型的对抗性强强性。此外,我们展示了数据印象在生成UAP时的功效。在几个基准数据集上进行的广泛的实验表明,在没有原始数据的情况下,使用原始数据来显示具有竞争性的印象。