Embedding of large but redundant data, such as images or text, in a hierarchy of lower-dimensional spaces is one of the key features of representation learning approaches, which nowadays provide state-of-the-art solutions to problems once believed hard or impossible to solve. In this work, in a plot twist with a strong meta aftertaste, we show how trained deep models are as redundant as the data they are optimized to process, and how it is therefore possible to use deep learning models to embed deep learning models. In particular, we show that it is possible to use representation learning to learn a fixed-size, low-dimensional embedding space of trained deep models and that such space can be explored by interpolation or optimization to attain ready-to-use models. We find that it is possible to learn an embedding space of multiple instances of the same architecture and of multiple architectures. We address image classification and neural representation of signals, showing how our embedding space can be learnt so as to capture the notions of performance and 3D shape, respectively. In the Multi-Architecture setting we also show how an embedding trained only on a subset of architectures can learn to generate already-trained instances of architectures it never sees instantiated at training time.
翻译:将大型但冗余的数据,如图像或文字,嵌入一个低维空间的层次,是代表性学习方法的关键特征之一,目前,这种方法为曾经认为难以或不可能解决的问题提供了最先进的解决办法。在这项工作中,我们用一个强有力的元后继自带曲曲折,展示了经过培训的深层模型如何与它们最优化的处理数据一样冗余,因此,如何利用深层学习模型嵌入深层学习模型。特别是,我们表明,可以使用代表性学习来学习一个由受过训练的深层模型构成的固定大小、低维深嵌入空间,而且这种空间可以通过内插或优化来探索,以达到现成的模型。我们发现,有可能学习同一架构和多个结构的多个实例的嵌入空间。我们处理图像分类和信号的神经表达,表明如何学习我们的嵌入空间,以便分别捕捉到性能和3D形状的概念。在多架构设置中,我们还展示了只有经过训练的嵌入空间才能在一组建筑结构中从不见即刻式实例。