Cognitive Cybersecurity for Artificial Intelligence: Guardrail Engineering with CCS-7
–arXiv.org Artificial Intelligence
Language models exhibit human-like cognitive vulnerabilities, such as emotional framing, that escape traditional behavioral alignment. We present CCS-7 (Cognitive Cybersecurity Suite), a taxonomy of seven vulnerabilities grounded in human cognitive security research. To establish a human benchmark, we ran a randomized controlled trial with 151 participants: a "Think First, Verify Always" (TFVA) lesson improved cognitive security by +7.9% overall. We then evaluated TFVA-style guardrails across 12,180 experiments on seven diverse language model architectures. Results reveal architecture-dependent risk patterns: some vulnerabilities (e.g., identity confusion) are almost fully mitigated, while others (e.g., source interference) exhibit escalating backfire, with error rates increasing by up to 135% in certain models. Humans, in contrast, show consistent moderate improvement. These findings reframe cognitive safety as a model-specific engineering problem: interventions effective in one architecture may fail, or actively harm, another, underscoring the need for architecture-aware cognitive safety testing before deployment.
arXiv.org Artificial Intelligence
Aug-15-2025
- Country:
- North America > United States
- Indiana > Marion County
- Indianapolis (0.04)
- New York (0.04)
- Indiana > Marion County
- North America > United States
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Strength High (1.00)
- Research Report
- Industry:
- Government > Military
- Cyberwarfare (0.61)
- Information Technology > Security & Privacy (1.00)
- Government > Military
- Technology: