FinDVer: Explainable Claim Verification over Long and Hybrid-Content Financial Documents
Zhao, Yilun, Long, Yitao, Jiang, Yuru, Wang, Chengye, Chen, Weiyuan, Liu, Hongjun, Zhang, Yiming, Tang, Xiangru, Zhao, Chen, Cohan, Arman
–arXiv.org Artificial Intelligence
We introduce FinDVer, a comprehensive benchmark specifically designed to evaluate the explainable claim verification capabilities of LLMs in the context of understanding and analyzing long, hybrid-content financial documents. FinDVer contains 2,400 expert-annotated examples, divided into three subsets: information extraction, numerical reasoning, and knowledge-intensive reasoning, each addressing common scenarios encountered in real-world financial contexts. We assess a broad spectrum of LLMs under long-context and RAG settings. Our results show that even the current best-performing system, GPT-4o, still lags behind human experts. We further provide in-depth analysis on long-context and RAG setting, Chain-of-Thought reasoning, and model reasoning errors, offering insights to drive future advancements. We believe that FinDVer can serve as a valuable benchmark for evaluating LLMs in claim verification over complex, expert-domain documents.
arXiv.org Artificial Intelligence
Nov-8-2024
- Country:
- North America > United States (0.93)
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Banking & Finance (1.00)
- Technology: