Few-shot learning aims to fast adapt a deep model from a few examples. While pre-training and meta-training can create deep models powerful for few-shot generalization, we find that pre-training and meta-training focuses respectively on cross-domain transferability and cross-task transferability, which restricts their data efficiency in the entangled settings of domain shift and task shift. We thus propose the Omni-Training framework to seamlessly bridge pre-training and meta-training for data-efficient few-shot learning. Our first contribution is a tri-flow Omni-Net architecture. Besides the joint representation flow, Omni-Net introduces two parallel flows for pre-training and meta-training, responsible for improving domain transferability and task transferability respectively. Omni-Net further coordinates the parallel flows by routing their representations via the joint-flow, enabling knowledge transfer across flows. Our second contribution is the Omni-Loss, which introduces a self-distillation strategy separately on the pre-training and meta-training objectives for boosting knowledge transfer throughout different training stages. Omni-Training is a general framework to accommodate many existing algorithms. Evaluations justify that our single framework consistently and clearly outperforms the individual state-of-the-art methods on both cross-task and cross-domain settings in a variety of classification, regression and reinforcement learning problems.
翻译:少见的学习旨在从几个例子中快速调整一个深层次的模型。虽然培训前和培训元培训可以创造出对少见的概括化有影响力的深层次模型,但我们发现,培训前和培训元培训分别侧重于跨域可转移性和跨任务可转移性,这限制了它们在交织的域转移和任务转移环境中的数据效率。因此,我们提议了Omni培训框架,以无缝地连接培训前和数据高效的微粒学习的元培训。我们的第一个贡献是三流的Omni-Net结构。除了联合代表流动外,Omni-Net还引入了两个平行的培训和元培训前流动,分别负责改进域可转移性和任务可转移性。Omni-Net通过联合流动来调整其表达方式,从而进一步协调平行流动的数据效率。我们的第二个贡献是Omni-Los,它提出了在培训前和元培训中分别促进不同培训阶段知识转让的自我提炼战略。Omni-Net是一个总的框架,它是一个总的框架,可以容纳许多现有的强化型的跨级的跨系统化的系统化的系统。评估,它可以清楚地解释。