Better Call Claude: Can LLMs Detect Changes of Writing Style?
Römisch, Johannes, Gorovaia, Svetlana, Halchynska, Mariia, Schmidt, Gleb, Yamshchikov, Ivan P.
–arXiv.org Artificial Intelligence
This article explores the zero-shot performance of state-of-the-art large language models (LLMs) on one of the most challenging tasks in authorship analysis: sentence-level style change detection. Benchmarking four LLMs on the official PAN~2024 and 2025 "Multi-Author Writing Style Analysis" datasets, we present several observations. First, state-of-the-art generative models are sensitive to variations in writing style - even at the granular level of individual sentences. Second, their accuracy establishes a challenging baseline for the task, outperforming suggested baselines of the PAN competition. Finally, we explore the influence of semantics on model predictions and present evidence suggesting that the latest generation of LLMs may be more sensitive to content-independent and purely stylistic signals than previously reported.
arXiv.org Artificial Intelligence
Sep-5-2025
- Country:
- Asia
- Europe
- Germany > Bavaria
- Lower Franconia > Würzburg (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Netherlands > Gelderland
- Nijmegen (0.04)
- Russia > Northwestern Federal District
- Leningrad Oblast > Saint Petersburg (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Germany > Bavaria
- North America
- Dominican Republic (0.04)
- United States
- Florida > Miami-Dade County
- Miami (0.04)
- New York > New York County
- New York City (0.04)
- Florida > Miami-Dade County
- Genre:
- Research Report > New Finding (0.93)
- Technology: