Baseline Defenses for Adversarial Attacks Against Aligned Language Models