Continual Test-Time Adaptation (CTTA) enables pre-trained models to adapt to continuously evolving domains. Existing methods have improved robustness but typically rely on fixed or batch-level thresholds, which cannot account for varying difficulty across classes and instances. This limitation is especially problematic in semantic segmentation, where each image requires dense, multi-class predictions. We propose an approach that adaptively adjusts pseudo labels to reflect the confidence distribution within each image and dynamically balances learning toward classes most affected by domain shifts. This fine-grained, class- and instance-aware adaptation produces more reliable supervision and mitigates error accumulation throughout continual adaptation. Extensive experiments across eight CTTA and TTA scenarios, including synthetic-to-real and long-term shifts, show that our method consistently outperforms state-of-the-art techniques, setting a new standard for semantic segmentation under evolving conditions.
翻译:持续测试时适应(CTTA)使预训练模型能够适应不断演变的领域。现有方法提升了鲁棒性,但通常依赖于固定或批处理级别的阈值,无法考虑不同类别和实例间的难度差异。这一局限在语义分割中尤为突出,因为每幅图像都需要密集的多类别预测。我们提出一种方法,能够自适应地调整伪标签以反映每幅图像内的置信度分布,并动态平衡学习过程,重点关注受域迁移影响最大的类别。这种细粒度、类别与实例感知的适应机制能够产生更可靠的监督信号,并在持续适应过程中减轻误差累积。在包括合成到真实场景及长期迁移在内的八种CTTA和TTA场景中进行的大量实验表明,我们的方法在性能上持续超越现有先进技术,为动态变化条件下的语义分割确立了新的基准。