Certifying LLM Safety against Adversarial Prompting