PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality
Yu, Byeongho, Lee, Changhun, Jin, Jungyu, Park, Eunhyeok
–arXiv.org Artificial Intelligence
To mitigate the hallucination problem in large language models, DoLa exploits early exit logits from the same model as a contrastive prior. However, we found that these early exit logits tend to be flat, low in magnitude, and fail to reflect meaningful contrasts. To address this, we propose PruneCD, a novel contrastive decoding method that constructs the amateur model via layer pruning rather than early exit. This design leads to more informative and well-aligned logits, enabling more effective contrastive decoding. Through qualitative and quantitative analyses, we demonstrate that PruneCD consistently improves factuality with minimal inference overhead, offering a robust and practical approach to mitigating hallucinations in LLMs.
arXiv.org Artificial Intelligence
Sep-24-2025
- Country:
- Asia
- Japan (0.04)
- South Korea > Gyeongsangbuk-do
- Pohang (0.04)
- North America > United States
- California (0.04)
- Asia
- Genre:
- Research Report (0.82)
- Technology: