Fraud-R1 : A Multi-Round Benchmark for Assessing the Robustness of LLM Against Augmented Fraud and Phishing Inducements
Yang, Shu, Zhu, Shenzhe, Wu, Zeyu, Wang, Keyu, Yao, Junchi, Wu, Junchao, Hu, Lijie, Li, Mengdi, Wong, Derek F., Wang, Di
–arXiv.org Artificial Intelligence
We introduce Fraud-R1, a benchmark designed to evaluate LLMs' ability to defend against internet fraud and phishing in dynamic, real-world scenarios. Fraud-R1 comprises 8,564 fraud cases sourced from phishing scams, fake job postings, social media, and news, categorized into 5 major fraud types. Unlike previous benchmarks, Fraud-R1 introduces a multi-round evaluation pipeline to assess LLMs' resistance to fraud at different stages, including credibility building, urgency creation, and emotional manipulation. Furthermore, we evaluate 15 LLMs under two settings: 1. Helpful-Assistant, where the LLM provides general decision-making assistance, and 2. Role-play, where the model assumes a specific persona, widely used in real-world agent-based interactions. Our evaluation reveals the significant challenges in defending against fraud and phishing inducement, especially in role-play settings and fake job postings. Additionally, we observe a substantial performance gap between Chinese and English, underscoring the need for improved multilingual fraud detection capabilities.
arXiv.org Artificial Intelligence
Feb-18-2025
- Country:
- Asia > China
- Hong Kong (0.27)
- North America > United States (1.00)
- Asia > China
- Genre:
- Research Report > New Finding (0.45)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Law Enforcement & Public Safety > Fraud (1.00)
- Technology: