🤖 AI Summary
To address catastrophic forgetting and error accumulation in continual test-time adaptation (CTTA) for semantic segmentation, this paper proposes an orthogonal projection subspace constraint and online prior aggregation framework. Methodologically: (i) it enforces model updates within an orthogonal projection subspace to explicitly preserve source-domain knowledge; (ii) it introduces an active image masking–based online prior aggregation mechanism to improve pseudo-label quality between teacher and student models; and (iii) it integrates parameter-efficient fine-tuning, online knowledge distillation, and mask-based augmentation to enable lightweight, dynamically pseudo-label–driven adaptation. Evaluated on multiple CTTA semantic segmentation benchmarks, the method achieves state-of-the-art performance—significantly outperforming existing approaches—while maintaining high accuracy, mitigating error propagation, and balancing adaptation effectiveness with computational efficiency.
📝 Abstract
Continual Test Time Adaptation (CTTA) is a task that requires a source pre-trained model to continually adapt to new scenarios with changing target distributions. Existing CTTA methods primarily focus on mitigating the challenges of catastrophic forgetting and error accumulation. Though there have been emerging methods based on forgetting adaptation with parameter-efficient fine-tuning, they still struggle to balance competitive performance and efficient model adaptation, particularly in complex tasks like semantic segmentation. In this paper, to tackle the above issues, we propose a novel pipeline, Orthogonal Projection Subspace to aggregate online Prior-knowledge, dubbed OoPk. Specifically, we first project a tuning subspace orthogonally which allows the model to adapt to new domains while preserving the knowledge integrity of the pre-trained source model to alleviate catastrophic forgetting. Then, we elaborate an online prior-knowledge aggregation strategy that employs an aggressive yet efficient image masking strategy to mimic potential target dynamism, enhancing the student model's domain adaptability. This further gradually ameliorates the teacher model's knowledge, ensuring high-quality pseudo labels and reducing error accumulation. We demonstrate our method with extensive experiments that surpass previous CTTA methods and achieve competitive performances across various continual TTA benchmarks in semantic segmentation tasks.