Better Call Claude: Can LLMs Detect Changes of Writing Style?

📅 2025-08-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses zero-shot sentence-level writing style change detection—a particularly challenging task in authorship analysis. We evaluate four state-of-the-art large language models (LLMs) in a zero-shot setting on the official PAN 2024/2025 datasets and design semantic-controlled experiments to disentangle content from stylistic signals. Results demonstrate that LLMs effectively capture content-invariant stylistic features, achieving significantly higher accuracy than the competition’s recommended baselines without any fine-tuning. Crucially, this work provides the first empirical evidence that current top-tier LLMs exhibit high sensitivity to fine-grained stylistic variations, enabling robust multi-author style boundary detection in an unsupervised, zero-shot manner. These findings establish a novel paradigm for LLM-driven unsupervised style analysis and deliver critical empirical support for deploying off-the-shelf LLMs in practical authorship attribution and style segmentation tasks.

Technology Category

Application Category

📝 Abstract
This article explores the zero-shot performance of state-of-the-art large language models (LLMs) on one of the most challenging tasks in authorship analysis: sentence-level style change detection. Benchmarking four LLMs on the official PAN~2024 and 2025 "Multi-Author Writing Style Analysis" datasets, we present several observations. First, state-of-the-art generative models are sensitive to variations in writing style - even at the granular level of individual sentences. Second, their accuracy establishes a challenging baseline for the task, outperforming suggested baselines of the PAN competition. Finally, we explore the influence of semantics on model predictions and present evidence suggesting that the latest generation of LLMs may be more sensitive to content-independent and purely stylistic signals than previously reported.
Problem

Research questions and friction points this paper is trying to address.

Detect sentence-level writing style changes using LLMs
Benchmark LLMs on multi-author style analysis datasets
Assess LLM sensitivity to content-independent stylistic signals
Innovation

Methods, ideas, or system contributions that make the work stand out.

Zero-shot LLMs detect sentence-level style changes
Generative models outperform PAN competition baselines
LLMs sensitive to content-independent stylistic signals
🔎 Similar Papers
No similar papers found.
J
Johannes Römisch
Center for Artificial Intelligence, Technical University of Applied Sciences Würzburg-Schweinfurt, Münzstraße 12, 97070, Würzburg, Germany
S
Svetlana Gorovaia
LEYA Lab, School of Computer Science, Physics and Technology, HSE University, 6, 25th Liniya, Vasilievsky Ostrov, 199004, St Petersburg, Russia
M
Mariia Halchynska
Center for Artificial Intelligence, Technical University of Applied Sciences Würzburg-Schweinfurt, Münzstraße 12, 97070, Würzburg, Germany
G
Gleb Schmidt
Humanities Lab, Faculty of Arts, Radboud University, Houtlaan 4, 6525 XZ, Nijmegen, Netherlands
Ivan P. Yamshchikov
Ivan P. Yamshchikov
Research Professor at CAIRO, THWS
natural language generationcomputational creativityempathetic aiethics of ai application