An unresolved problem in Deep Learning is the ability of neural networks to cope with domain shifts during test-time, imposed by commonly fixing network parameters after training. Our proposed method Meta Test-Time Training (MT3), however, breaks this paradigm and enables adaption at test-time. We combine meta-learning, self-supervision and test-time training to learn to adapt to unseen test distributions. By minimizing the self-supervised loss, we learn task-specific model parameters for different tasks. A meta-model is optimized such that its adaption to the different task-specific models leads to higher performance on those tasks. During test-time a single unlabeled image is sufficient to adapt the meta-model parameters. This is achieved by minimizing only the self-supervised loss component resulting in a better prediction for that image. Our approach significantly improves the state-of-the-art results on the CIFAR-10-Corrupted image classification benchmark. Our implementation is available on GitHub.
翻译:深层学习中的一个未决问题是神经网络在测试期间应对测试时域变换的能力,这是由培训后通常确定网络参数所强加的。然而,我们提出的方法Meta测试-时间培训(MT3)打破了这一模式,能够在测试时适应。我们把元学习、自我监督以及测试-时间培训结合起来,学习适应无形的测试分布。通过尽量减少自我监督的损失,我们学习了不同任务的具体任务模式参数。一个元模型得到优化,使其适应不同任务模式导致这些任务的更高性能。在测试时,一个单一的无标签图像足以适应元模型参数。实现这一点的办法是,只尽量减少导致更好地预测该图像的自我监督损失部分。我们的方法大大改进了CIFAR-10校正的图像分类基准的最新结果。我们在GitHub上可以使用。