A Granular Study of Safety Pretraining under Model Abliteration
Agnihotri, Shashank, Jakubassa, Jonas, Dey, Priyam, Goyal, Sachin, Schiele, Bernt, Radhakrishnan, Venkatesh Babu, Keuper, Margret
–arXiv.org Artificial Intelligence
Open-weight LLMs can be modified at inference time with simple activation edits, which raises a practical question for safety: do common safety interventions like refusal training or metatag training survive such edits? We study model abliteration, a lightweight projection technique designed to remove refusal-sensitive directions, and conduct a controlled evaluation across a granular sequence of Safety Pretraining checkpoints for SmolLM2-1.7B, alongside widely used open baselines. For each of 20 systems, original and abliterated, we issue 100 prompts with balanced harmful and harmless cases, classify responses as **Refusal** or **Non-Refusal** using multiple judges, and validate judge fidelity on a small human-labeled subset. We also probe whether models can identify refusal in their own outputs. Our study produces a checkpoint-level characterization of which data-centric safety components remain robust under abliteration, quantifies how judge selection influences evaluation outcomes, and outlines a practical protocol for integrating inference-time edits into safety assessments. Code: https://github.com/shashankskagnihotri/safety_pretraining.
arXiv.org Artificial Intelligence
Oct-6-2025
- Country:
- Asia > India
- Europe
- Germany
- Baden-Württemberg (0.04)
- Saarland (0.04)
- Latvia > Lubāna Municipality
- Lubāna (0.04)
- Germany
- North America > United States
- Florida > Hillsborough County > University (0.04)
- Genre:
- Research Report > Experimental Study (1.00)
- Technology: