Causal-Tune: Mining Causal Factors from Vision Foundation Models for Domain Generalized Semantic Segmentation

Authors

  • Yin Zhang School of Instrument Science and Engineering, Harbin Institute of Technology, Harbin, China Computer Vision Center, Universitat Autònoma de Barcelona, Barcelona, Spain
  • Yongqiang Zhang College of Computer Science, Inner Mongolia University, Inner Mongolia, China
  • Yaoyue Zheng Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Shaanxi, China Computer Vision Center, Universitat Autònoma de Barcelona, Barcelona, Spain
  • Bogdan Raducanu Computer Vision Center, Universitat Autònoma de Barcelona, Barcelona, Spain
  • Dan Liu School of Instrument Science and Engineering, Harbin Institute of Technology, Harbin, China

DOI:

https://doi.org/10.1609/aaai.v40i15.38290

Abstract

Fine-tuning Vision Foundation Models (VFMs) with a small number of parameters has shown remarkable performance in Domain Generalized Semantic Segmentation (DGSS). Most existing works either train lightweight adapters or refine intermediate features to achieve better generalization on unseen domains. However, they both overlook the fact that long-term pre-trained VFMs often exhibit artifacts, which hinder the utilization of valuable representations and ultimately degrade DGSS performance. Inspired by causal mechanisms, we observe that these artifacts are associated with non-causal factors, which usually reside in the low- and high-frequency components of the VFM spectrum. In this paper, we explicitly examine the causal and non-causal factors of features within VFMs for DGSS, and propose a simple yet effective method to identify and disentangle them, enabling more robust domain generalization. Specifically, we propose Causal-Tune, a novel fine-tuning strategy designed to extract causal factors and suppress non-causal ones from the features of VFMs. First, we extract the frequency spectrum of features from each layer using the Discrete Cosine Transform (DCT). A Gaussian band-pass filter is then applied to separate the spectrum into causal and non-causal components. To further refine the causal components, we introduce a set of causal-aware learnable tokens that operate in the frequency domain, while the non-causal components are discarded. Finally, refined features are transformed back into the spatial domain via inverse DCT and passed to the next layer. Extensive experiments conducted on various cross-domain tasks demonstrate the effectiveness of Causal-Tune. In particular, our method achieves superior performance under adverse weather conditions, improving +4.8% mIoU over the baseline in snow conditions.

Published

2026-03-14

How to Cite

Zhang, Y., Zhang, Y., Zheng, Y., Raducanu, B., & Liu, D. (2026). Causal-Tune: Mining Causal Factors from Vision Foundation Models for Domain Generalized Semantic Segmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 40(15), 12916–12924. https://doi.org/10.1609/aaai.v40i15.38290

Issue

Section

AAAI Technical Track on Computer Vision XII