Self-Supervised Learning (SSL) has been shown to learn useful and information-preserving representations. Neural Networks (NNs) are widely applied, yet their weight space is still not fully understood. Therefore, we propose to use SSL to learn hyper-representations of the weights of populations of NNs. To that end, we introduce domain specific data augmentations and an adapted attention architecture. Our empirical evaluation demonstrates that self-supervised representation learning in this domain is able to recover diverse NN model characteristics. Further, we show that the proposed learned representations outperform prior work for predicting hyper-parameters, test accuracy, and generalization gap as well as transfer to out-of-distribution settings.
翻译:自我监督学习(SSL)被证明学习了有用和信息保存的表达方式。神经网络(NNS)被广泛应用,但其重量空间仍然不完全理解。因此,我们提议使用SSL学习非军事网络人口重量的超强代表比例。为此,我们引入了特定领域的数据增强和调整关注结构。我们的经验评估表明,在这一领域自我监督的表述学习能够恢复非军事网络模式的多种特点。此外,我们表明,拟议的学习表达方式超过了预测超参数、测试准确度和一般化差距以及转移到分配之外环境的先前工作。