Making LLMs Vulnerable to Prompt Injection via Poisoning Alignment