LLaVul: A Multimodal LLM for Interpretable Vulnerability Reasoning about Source Code