A major problem of deep neural networks for image classification is their vulnerability to domain changes at test-time. Recent methods have proposed to address this problem with test-time training (TTT), where a two-branch model is trained to learn a main classification task and also a self-supervised task used to perform test-time adaptation. However, these techniques require defining a proxy task specific to the target application. To tackle this limitation, we propose TTTFlow: a Y-shaped architecture using an unsupervised head based on Normalizing Flows to learn the normal distribution of latent features and detect domain shifts in test examples. At inference, keeping the unsupervised head fixed, we adapt the model to domain-shifted examples by maximizing the log likelihood of the Normalizing Flow. Our results show that our method can significantly improve the accuracy with respect to previous works.
翻译:用于图像分类的深神经网络的一个主要问题是它们易受测试时域变化的影响。最近的一些方法提议通过测试时间培训(TTT)来解决这一问题,在测试时间培训中,一个两部门模式经过培训,学习主要的分类任务和用于测试时间适应的自监督任务。然而,这些技术要求确定目标应用程序特有的代理任务。为了应对这一限制,我们建议TTFlow:一个Y型结构,使用基于正常流程的未受监督的头来学习潜在特征的正常分布,并在测试示例中检测域变化。推断,保持未受监督的头部固定,我们通过最大限度地增加正常流程的日志可能性,将模型调整为域变实例。我们的结果显示,我们的方法可以大大改进以往工程的准确性。