LangDA: Building Context-Awareness via Language for Domain Adaptive Semantic Segmentation
Liu, Chang, Balaji, Bavesh, Hossain, Saad, Thomas, C, Lai, Kwei-Herng, Vemulapalli, Raviteja, Wong, Alexander, Rambhatla, Sirisha
Unsupervised domain adaptation for semantic segmentation (DASS) aims to transfer knowledge from a label-rich source domain to a target domain with no labels. Two key approaches in DASS are (1) vision-only approaches using masking or multi-resolution crops, and (2) language-based approaches that use generic class-wise prompts informed by target domain (e.g. "a {snowy} photo of a {class}"). However, the former is susceptible to noisy pseudo-labels that are biased to the source domain. The latter does not fully capture the intricate spatial relationships of objects -- key for dense prediction tasks. To this end, we propose LangDA. LangDA addresses these challenges by, first, learning contextual relationships between objects via VLM-generated scene descriptions (e.g. "a pedestrian is on the sidewalk, and the street is lined with buildings."). Second, LangDA aligns the entire image features with text representation of this context-aware scene caption and learns generalized representations via text. With this, LangDA sets the new state-of-the-art across three DASS benchmarks, outperforming existing methods by 2.6%, 1.4% and 3.9%.
Mar-16-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- Netherlands > North Holland
- Amsterdam (0.04)
- Romania > Sud - Muntenia Development Region
- Giurgiu County > Giurgiu (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- Netherlands > North Holland
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.46)
- Technology: