Recently, methods have been developed to accurately predict the testing performance of a Deep Neural Network (DNN) on a particular task, given statistics of its underlying topological structure. However, further leveraging this newly found insight for practical applications is intractable due to the high computational cost in terms of time and memory. In this work, we define a new class of topological features that accurately characterize the progress of learning while being quick to compute during running time. Additionally, our proposed topological features are readily equipped for backpropagation, meaning that they can be incorporated in end-to-end training. Our newly developed practical topological characterization of DNNs allows for an additional set of applications. We first show we can predict the performance of a DNN without a testing set and without the need for high-performance computing. We also demonstrate our topological characterization of DNNs is effective in estimating task similarity. Lastly, we show we can induce learning in DNNs by actively constraining the DNN's topological structure. This opens up new avenues in constricting the underlying structure of DNNs in a meta-learning framework.
翻译:最近,根据深神经网络的基本地形结构的统计,开发了一些方法,准确预测深神经网络在特定任务上的测试性能。然而,由于时间和记忆方面的计算成本很高,进一步利用这一新发现的实际应用的洞察力很难。在这项工作中,我们定义了一种新的地形特征类别,准确描述学习进展,同时在运行期间快速进行计算。此外,我们提议的地形特征很容易用于后向分析,这意味着这些特征可以纳入端到端培训中。我们新开发的DNN的实用地形特征允许增加一套应用程序。我们首先显示,我们可以预测DNN的性能,无需测试,不需要高性能计算。我们还表明,DNN的地形特征在估算任务相似性方面是有效的。最后,我们表明,通过积极限制DNN的表性结构,我们可以在DNN的顶端学习中进行学习。这打开了将DNN的基本结构压缩在元学习框架中的新途径。