SLM Meets LLM: Balancing Latency, Interpretability and Consistency in Hallucination Detection
Hu, Mengya, Xu, Rui, Lei, Deren, Li, Yaxi, Wang, Mingyu, Ching, Emily, Kamal, Eslam, Deng, Alex
–arXiv.org Artificial Intelligence
Large language models (LLMs) are highly capable but face latency challenges in real-time applications, such as conducting online hallucination detection. To overcome this issue, we propose a novel framework that leverages a small language model (SLM) classifier for initial detection, followed by a LLM as constrained reasoner to generate detailed explanations for detected hallucinated content. This study optimizes the real-time interpretable hallucination detection by introducing effective prompting techniques that align LLM-generated explanations with SLM decisions. Empirical experiment results demonstrate its effectiveness, thereby enhancing the overall user experience.
arXiv.org Artificial Intelligence
Aug-22-2024
- Country:
- Asia > Singapore (0.04)
- North America > United States (0.05)
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Leisure & Entertainment (0.47)
- Media (0.47)
- Technology: