DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios Junchao Wu1 Derek F. Wong 1 Shu Yang 1
–Neural Information Processing Systems
Detecting text generated by large language models (LLMs) is of great recent interest. With zero-shot methods like DetectGPT, detection capabilities have reached impressive levels. However, the reliability of existing detectors in real-world applications remains underexplored. In this study, we present a new benchmark, DetectRL, highlighting that even state-of-the-art (SOTA) detection techniques still underperformed in this task. We collected human-written datasets from domains where LLMs are particularly prone to misuse. Using popular LLMs, we generated data that better aligns with real-world applications. Unlike previous studies, we employed heuristic rules to create adversarial LLM-generated text, simulating various prompts usages, human revisions like word substitutions, and writing noises like spelling mistakes. Our development of DetectRL reveals the strengths and limitations of current SOTA detectors. More importantly, we analyzed the potential impact of writing styles, model types, attack methods, the text lengths, and real-world human writing factors on different types of detectors.
Neural Information Processing Systems
Mar-27-2025, 02:16:15 GMT
- Country:
- Asia (1.00)
- Europe
- Austria > Vienna (0.14)
- Middle East > Malta (0.14)
- North America
- Canada > Quebec (0.14)
- United States
- California > San Francisco County
- San Francisco (0.14)
- Hawaii (0.14)
- California > San Francisco County
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Technology: