LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples