Detecting Machine-Generated Long-Form Content with Latent-Space Variables
Tian, Yufei, Pan, Zeyu, Peng, Nanyun
–arXiv.org Artificial Intelligence
The increasing capability of large language models (LLMs) to generate fluent long-form texts is presenting new challenges in distinguishing machine-generated outputs from human-written ones, which is crucial for ensuring authenticity and trustworthiness of expressions. Existing zero-shot detectors primarily focus on token-level distributions, which are vulnerable to real-world domain shifts, including different prompting and decoding strategies, and adversarial attacks. We propose a more robust method that incorporates abstract elements, such as event transitions, as key deciding factors to detect machine versus human texts by training a latent-space model on sequences of events or topics derived from human-written texts. In three different domains, machine-generated texts, which are originally inseparable from human texts on the token level, can be better distinguished with our latent-space model, leading to a 31% improvement over strong baselines such as DetectGPT. Our analysis further reveals that, unlike humans, modern LLMs like GPT-4 generate event triggers and their transitions differently, an inherent disparity that helps our method to robustly detect machine-generated texts.
arXiv.org Artificial Intelligence
Oct-4-2024
- Country:
- Asia (1.00)
- Europe (1.00)
- North America > United States (0.67)
- Genre:
- Research Report (0.82)
- Industry:
- Information Technology > Security & Privacy (0.48)
- Media > Film (0.46)
- Technology: