LRQ-Fact: LLM-Generated Relevant Questions for Multimodal Fact-Checking
Beigi, Alimohammad, Jiang, Bohan, Li, Dawei, Kumarage, Tharindu, Tan, Zhen, Shaeri, Pouya, Liu, Huan
–arXiv.org Artificial Intelligence
Human fact-checkers have specialized domain knowledge that allows them to formulate precise questions to verify information accuracy. However, this expert-driven approach is labor-intensive and is not scalable, especially when dealing with complex multimodal misinformation. In this paper, we propose a fully-automated framework, LRQ-Fact, for multimodal fact-checking. Firstly, the framework leverages Vision-Language Models (VLMs) and Large Language Models (LLMs) to generate comprehensive questions and answers for probing multimodal content. Next, a rule-based decision-maker module evaluates both the original content and the generated questions and answers to assess the overall veracity. Extensive experiments on two benchmarks show that LRQ-Fact improves detection accuracy for multimodal misinformation. Moreover, we evaluate its generalizability across different model backbones, offering valuable insights for further refinement.
arXiv.org Artificial Intelligence
Oct-6-2024
- Country:
- North America > United States > Arizona (0.14)
- Genre:
- Research Report (0.82)
- Industry:
- Government (1.00)
- Health & Medicine > Therapeutic Area (1.00)
- Information Technology > Security & Privacy (1.00)
- Media > News (1.00)
- Technology: