UniGen: Universal Domain Generalization for Sentiment Classification via Zero-shot Dataset Generation
Choi, Juhwan, Kim, Yeonghwa, Yu, Seunguk, Yun, JungMin, Kim, YoungBin
–arXiv.org Artificial Intelligence
Although pre-trained language models have exhibited great flexibility and versatility with prompt-based few-shot learning, they suffer from the extensive parameter size and limited applicability for inference. Recent studies have suggested that PLMs be used as dataset generators and a tiny task-specific model be trained to achieve efficient inference. However, their applicability to various domains is limited because they tend to generate domain-specific datasets. In this work, we propose a novel approach to universal domain generalization that generates a dataset regardless of the target domain. This allows for generalization of the tiny task model to any domain that shares the label space, thus enhancing the real-world applicability of the dataset generation paradigm. Our experiments indicate that the proposed method accomplishes generalizability across various domains while using a parameter set that is orders of magnitude smaller than PLMs.
arXiv.org Artificial Intelligence
May-2-2024
- Country:
- Asia > South Korea > Seoul > Seoul (0.04)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Leisure & Entertainment (0.68)
- Media (0.46)
- Technology: