Discovering Forbidden Topics in Language Models
Rager, Can, Wendler, Chris, Gandikota, Rohit, Bau, David
–arXiv.org Artificial Intelligence
Refusal discovery is the task of identifying the full set of topics that a language model refuses to discuss. We introduce this new problem setting and develop a refusal discovery method, Iterated Prefill Crawler (IPC), that uses token prefilling to find forbidden topics. We benchmark IPC on Tulu-3-8B, an open-source model with public safety tuning data. Our crawler manages to retrieve 31 out of 36 topics within a budget of 1000 prompts. Next, we scale the crawler to a frontier model using the prefilling option of Claude-Haiku. Finally, we crawl three widely used open-weight models: Llama-3.3-70B and two of its variants finetuned for reasoning: DeepSeek-R1-70B and Perplexity-R1-1776-70B. DeepSeek-R1-70B reveals patterns consistent with censorship tuning: The model exhibits "thought suppression" behavior that indicates memorization of CCP-aligned responses. Although Perplexity-R1-1776-70B is robust to censorship, IPC elicits CCP-aligned refusals answers in the quantized model. Our findings highlight the critical need for refusal discovery methods to detect biases, boundaries, and alignment failures of AI systems.
arXiv.org Artificial Intelligence
Jun-12-2025
- Country:
- Asia
- Europe > Ukraine (0.04)
- North America > United States (0.04)
- Oceania > New Zealand (0.04)
- Genre:
- Personal > Interview (0.93)
- Research Report > New Finding (0.87)
- Industry:
- Government
- Military (0.93)
- Regional Government > Asia Government
- China Government (1.00)
- Health & Medicine
- Pharmaceuticals & Biotechnology (0.92)
- Therapeutic Area > Psychiatry/Psychology (0.67)
- Information Technology > Security & Privacy (1.00)
- Law
- Civil Rights & Constitutional Law (1.00)
- Criminal Law (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- Government
- Technology: