Fine-tuning multilingual language models in Twitter/X sentiment analysis: a study on Eastern-European V4 languages
Filip, Tomáš, Pavlíček, Martin, Sosík, Petr
–arXiv.org Artificial Intelligence
The aspect-based sentiment analysis (ABSA) is a standard NLP task with numerous approaches and benchmarks, where large language models (LLM) represent the current state-of-the-art. We focus on ABSA subtasks based on Twitter/X data in underrepresented languages. On such narrow tasks, small tuned language models can often outperform universal large ones, providing available and cheap solutions. We fine-tune several LLMs (BERT, BERTweet, Llama2, Llama3, Mistral) for classification of sentiment towards Russia and Ukraine in the context of the ongoing military conflict. The training/testing dataset was obtained from the academic API from Twitter/X during 2023, narrowed to the languages of the V4 countries (Czech Republic, Slovakia, Poland, Hungary). Then we measure their performance under a variety of settings including translations, sentiment targets, in-context learning and more, using GPT4 as a reference model. We document several interesting phenomena demonstrating, among others, that some models are much better fine-tunable on multilingual Twitter tasks than others, and that they can reach the SOTA level with a very small training set. Finally we identify combinations of settings providing the best results.
arXiv.org Artificial Intelligence
Aug-4-2024
- Country:
- Asia > Russia (0.53)
- Europe
- Czechia > Moravian-Silesian Region
- Ostrava (0.04)
- Hungary (0.24)
- Sweden (0.04)
- Ukraine (0.39)
- Finland > Uusimaa
- Helsinki (0.06)
- Russia (0.29)
- Slovakia (0.24)
- Poland (0.25)
- Netherlands > North Holland
- Amsterdam (0.04)
- Czechia > Moravian-Silesian Region
- North America > United States
- Minnesota > Hennepin County > Minneapolis (0.14)
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Government > Military (0.66)
- Information Technology > Security & Privacy (0.46)
- Technology: