SF(DA)$^2$: Source-free Domain Adaptation Through the Lens of Data Augmentation
Hwang, Uiwon, Lee, Jonghyun, Shin, Juhyeon, Yoon, Sungroh
–arXiv.org Artificial Intelligence
In the face of the deep learning model's vulnerability to domain shift, source-free domain adaptation (SFDA) methods have been proposed to adapt models to new, unseen target domains without requiring access to source domain data. Although the potential benefits of applying data augmentation to SFDA are attractive, several challenges arise such as the dependence on prior knowledge of class-preserving transformations and the increase in memory and computational requirements. We construct an augmentation graph in the feature space of the pretrained model using the neighbor relationships between target features and propose spectral neighborhood clustering to identify partitions in the prediction space. Furthermore, we propose implicit feature augmentation and feature disentanglement as regularization loss functions that effectively utilize class semantic information within the feature space. These regularizers simulate the inclusion of an unlimited number of augmented target features into the augmentation graph while minimizing computational and memory demands. Our method shows superior adaptation performance in SFDA scenarios, including 2D image and 3D point cloud datasets and a highly imbalanced dataset. In recent years, deep learning has achieved significant advancements and is widely explored for realworld applications. However, the performance of deep learning models can significantly deteriorate when deployed on unlabeled target domains, which differ from the source domain where the training data was collected. This domain shift poses a challenge for applying deep learning models in practical scenarios.
arXiv.org Artificial Intelligence
Mar-16-2024