Model-Agnostic Sentiment Distribution Stability Analysis for Robust LLM-Generated Texts Detection
Li, Siyuan, Lin, Xi, Li, Guangyan, Liu, Zehao, Wulianghai, Aodu, Ding, Li, Wu, Jun, Li, Jianhua
–arXiv.org Artificial Intelligence
The rapid advancement of large language models (LLMs) has resulted in increasingly sophisticated AI-generated content, posing significant challenges in distinguishing LLM-generated text from human-written language. Existing detection methods, primarily based on lexical heuristics or fine-tuned classifiers, often suffer from limited generalizability and are vulnerable to paraphrasing, adversarial perturbations, and cross-domain shifts. In this work, we propose SentiDetect, a model-agnostic framework for detecting LLM-generated text by analyzing the divergence in sentiment distribution stability. Our method is motivated by the empirical observation that LLM outputs tend to exhibit emotionally consistent patterns, whereas human-written texts display greater emotional variability. To capture this phenomenon, we define two complementary metrics: sentiment distribution consistency and sentiment distribution preservation, which quantify stability under sentiment-altering and semantic-preserving transformations. We evaluate SentiDetect on five diverse datasets and a range of advanced LLMs,including Gemini-1.5-Pro, Claude-3, GPT-4-0613, and LLaMa-3.3. Experimental results demonstrate its superiority over state-of-the-art baselines, with over 16% and 11% F1 score improvements on Gemini-1.5-Pro and GPT-4-0613, respectively. Moreover, SentiDetect also shows greater robustness to paraphrasing, adversarial attacks, and text length variations, outperforming existing detectors in challenging scenarios.
arXiv.org Artificial Intelligence
Aug-12-2025
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Education > Curriculum
- Subject-Specific Education (0.48)
- Information Technology > Security & Privacy (0.88)
- Media (0.68)
- Education > Curriculum
- Technology: