Delta - Contrastive Decoding Mitigates Text Hallucinations in Large Language Models
Huang, Cheng Peng, Chen, Hao-Yuan
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) have demonstrated remarkable capabilities in natural language processing tasks. Still, they are prone to generating hallucinations--factually incorrect or fabricated content that can undermine their reliability, especially in high-stakes domains such as healthcare and legal advisory. In response to this challenge, we propose Delta, a novel inference-time approach that leverages contrastive decoding to mitigate hallucinations without requiring model retraining or additional training data. Delta works by randomly masking portions of the input prompt, then contrasting the original and masked output distribution generated by the model, effectively mitigating hallucinations through inferenceonly computations. Delta was evaluated on context-rich QA benchmarks like SQuAD v1.1 and v2, achieving around 3 and 6 percentage points of improvement, respectively. It also showed gains of 7 and 2 percentage points on TriviaQA and Natural Question under-sampling decoding. Delta improved SQuAD v2's noanswer exact match by over ten percentage points.
arXiv.org Artificial Intelligence
Feb-9-2025
- Country:
- Asia
- Europe
- Monaco (0.04)
- United Kingdom > England
- Greater London > London (0.04)
- North America
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Ohio > Delaware County
- Delaware (0.04)
- Texas > Travis County
- Austin (0.04)
- Minnesota > Hennepin County
- Mexico > Mexico City
- Genre:
- Research Report > New Finding (0.68)
- Technology: