Test-Time-Training (TTT) is an approach to cope with out-of-distribution (OOD) data by adapting a trained model to distribution shifts occurring at test-time. We propose to perform this adaptation via Activation Matching (ActMAD): We analyze activations of the model and align activation statistics of the OOD test data to those of the training data. In contrast to existing methods, which model the distribution of entire channels in the ultimate layer of the feature extractor, we model the distribution of each feature in multiple layers across the network. This results in a more fine-grained supervision and makes ActMAD attain state of the art performance on CIFAR-100C and Imagenet-C. ActMAD is also architecture- and task-agnostic, which lets us go beyond image classification, and score 15.4% improvement over previous approaches when evaluating a KITTI-trained object detector on KITTI-Fog. Our experiments highlight that ActMAD can be applied to online adaptation in realistic scenarios, requiring little data to attain its full performance.
翻译:测试时间训练(TTT)是应对测试时发生的分布偏移的一种方法。我们提出通过激活匹配(ActMAD)来实现此自适应:我们分析模型的激活并将OOD测试数据的激活统计信息与训练数据的激活统计信息进行对齐。与现有方法不同,现有方法模型化特征提取器的最终层中整个通道的分布,我们模型化网络中多层的每个特征的分布。这导致了更细粒度的监督,并使ActMAD在CIFAR-100C和Imagenet-C上达到了最新水平。ActMAD还是架构和任务无关的,可以超越图像分类,评估KITTI训练的物体检测器在KITTI-Fog上得分15.4%的改进。我们的实验强调ActMAD可以应用于现实场景中的在线自适应,需要很少的数据才能实现其完整性能。