Aligned Probing: Relating Toxic Behavior and Model Internals
Waldis, Andreas, Gautam, Vagrant, Lauscher, Anne, Klakow, Dietrich, Gurevych, Iryna
–arXiv.org Artificial Intelligence
We introduce aligned probing, a novel interpretability framework that aligns the behavior of language models (LMs), based on their outputs, and their internal representations (internals). Using this framework, we examine over 20 OLMo, Llama, and Mistral models, bridging behavioral and internal perspectives for toxicity for the first time. Our results show that LMs strongly encode information about the toxicity level of inputs and subsequent outputs, particularly in lower layers. Focusing on how unique LMs differ offers both correlative and causal evidence that they generate less toxic output when strongly encoding information about the input toxicity. We also highlight the heterogeneity of toxicity, as model behavior and internals vary across unique attributes such as Threat. Finally, four case studies analyzing detoxification, multi-prompt evaluations, model quantization, and pre-training dynamics underline the practical impact of aligned probing with further concrete insights. Our findings contribute to a more holistic understanding of LMs, both within and beyond the context of toxicity.
arXiv.org Artificial Intelligence
Mar-17-2025
- Country:
- Asia > Middle East
- UAE (0.14)
- Europe
- Austria > Vienna (0.14)
- Middle East > Malta (0.14)
- North America
- Mexico > Mexico City (0.14)
- United States
- Asia > Middle East
- Genre:
- Research Report > New Finding (1.00)
- Technology: