Safety Evaluation of DeepSeek Models in Chinese Contexts
Zhang, Wenjing, Lei, Xuejiao, Liu, Zhaoxiang, Wang, Ning, Long, Zhenhong, Yang, Peijun, Zhao, Jiaojiao, Hua, Minjie, Ma, Chaoyang, Wang, Kai, Lian, Shiguo
–arXiv.org Artificial Intelligence
Recently, the DeepSeek series of models, leveraging their exceptional reasoning capabilities and open-source strategy, is reshaping the global AI landscape. Despite these advantages, they exhibit significant safety deficiencies. Research conducted by Robust Intelligence, a subsidiary of Cisco, in collaboration with the University of Pennsylvania, revealed that DeepSeek-R1 has a 100\% attack success rate when processing harmful prompts. Additionally, multiple safety companies and research institutions have confirmed critical safety vulnerabilities in this model. As models demonstrating robust performance in Chinese and English, DeepSeek models require equally crucial safety assessments in both language contexts. However, current research has predominantly focused on safety evaluations in English environments, leaving a gap in comprehensive assessments of their safety performance in Chinese contexts. In response to this gap, this study introduces CHiSafetyBench, a Chinese-specific safety evaluation benchmark. This benchmark systematically evaluates the safety of DeepSeek-R1 and DeepSeek-V3 in Chinese contexts, revealing their performance across safety categories. The experimental results quantify the deficiencies of these two models in Chinese contexts, providing key insights for subsequent improvements.
arXiv.org Artificial Intelligence
Feb-16-2025
- Country:
- Asia > China (0.29)
- North America > United States
- Pennsylvania (0.25)
- Genre:
- Research Report (0.64)
- Technology: