JailbreakBench: An Open Robustness Benchmark for Jailbreaking Large Language Models