OutSafe-Bench: A Benchmark for Multimodal Offensive Content Detection in Large Language Models