We formulate an asymmetric (or non-commutative) distance between tasks based on Fisher Information Matrices, called Fisher task distance. This distance represents the complexity of transferring the knowledge from one task to another. We provide a proof of consistency for our distance through theorems and experiments on various classification tasks from MNIST, CIFAR-10, CIFAR-100, ImageNet, and Taskonomy datasets. Next, we construct an online neural architecture search framework using the Fisher task distance, in which we have access to the past learned tasks. By using the Fisher task distance, we can identify the closest learned tasks to the target task, and utilize the knowledge learned from these related tasks for the target task. Here, we show how the proposed distance between a target task and a set of learned tasks can be used to reduce the neural architecture search space for the target task. The complexity reduction in search space for task-specific architectures is achieved by building on the optimized architectures for similar tasks instead of doing a full search and without using this side information. Experimental results for tasks in MNIST, CIFAR-10, CIFAR-100, ImageNet datasets demonstrate the efficacy of the proposed approach and its improvements, in terms of the performance and the number of parameters, over other gradient-based search methods, such as ENAS, DARTS, PC-DARTS.
翻译:我们根据渔业信息矩阵制定任务之间的不对称(或非互换性)距离,称为渔业任务距离。这种距离代表了将知识从一个任务转移到另一个任务的复杂性。我们通过理论和对来自MNIST、CIFAR-10、CIFAR-100、图像网和任务数据集的各种分类任务进行实验,为我们的距离提供了证据。接下来,我们利用Fisher任务距离建立一个在线神经结构搜索框架,我们可以进入过去学到的任务。通过利用渔业任务距离,我们可以确定最接近目标任务的任务,并利用从这些相关任务中学到的知识来完成目标任务。我们在这里展示了如何利用目标任务与一组学习任务之间的拟议距离来减少目标任务的神经结构搜索空间。具体任务结构搜索空间的复杂程度是通过建立类似任务的最佳结构来实现的,而不是完全搜索和不使用这一侧面信息。我们可以通过MNIST、CIFAR-10、CIFAR-100、ART-EN-AS的图像网络方法的实验结果,展示了拟议搜索方式的效能,例如搜索方式的搜索方式,以及AS-D的升级方法。