Tell Me Why: Explainable Public Health Fact-Checking with Large Language Models
Zarharan, Majid, Wullschleger, Pascal, Kia, Babak Behkam, Pilehvar, Mohammad Taher, Foster, Jennifer
–arXiv.org Artificial Intelligence
This paper presents a comprehensive analysis of explainable fact-checking through a series of experiments, focusing on the ability of large language models to verify public health claims and provide explanations or justifications for their veracity assessments. We examine the effectiveness of zero/few-shot prompting and parameter-efficient fine-tuning across various open and closed-source models, examining their performance in both isolated and joint tasks of veracity prediction and explanation generation. Importantly, we employ a dual evaluation approach comprising previously established automatic metrics and a novel set of criteria through human evaluation. Our automatic evaluation indicates that, within the zero-shot scenario, GPT-4 emerges as the standout performer, but in few-shot and parameter-efficient fine-tuning contexts, open-source models demonstrate their capacity to not only bridge the performance gap but, in some instances, surpass GPT-4. Human evaluation reveals yet more nuance as well as indicating potential problems with the gold explanations.
arXiv.org Artificial Intelligence
May-15-2024
- Country:
- Africa > South Africa
- Gauteng (0.04)
- Asia
- Indonesia (0.04)
- Middle East
- Iran (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- Belgium > Brussels-Capital Region
- North America
- Puerto Rico (0.05)
- United States
- Louisiana > Orleans Parish
- New Orleans (0.04)
- New York > Suffolk County
- Stony Brook (0.04)
- Pennsylvania (0.04)
- Texas (0.04)
- Louisiana > Orleans Parish
- Oceania > Australia (0.04)
- Africa > South Africa
- Genre:
- Research Report (1.00)
- Industry:
- Technology: