CoTEVer: Chain of Thought Prompting Annotation Toolkit for Explanation Verification
Kim, Seungone, Joo, Se June, Jang, Yul, Chae, Hyungjoo, Yeo, Jinyoung
–arXiv.org Artificial Intelligence
Chain-of-thought (CoT) prompting enables large language models (LLMs) to solve complex reasoning tasks by generating an explanation before the final prediction. Despite it's promising ability, a critical downside of CoT prompting is that the performance is greatly affected by the factuality of the generated explanation. To improve the correctness of the explanations, fine-tuning language models with explanation data is needed. However, there exists only a few datasets that can be used for such approaches, and no data collection tool for building them. Thus, we introduce CoTEVer, a tool-kit for annotating the factual correctness of generated explanations and collecting revision data of wrong explanations. Figure 1: Example of Explanation Verification and Answer Furthermore, we suggest several use cases Verification of GPT-3's output. Explanation Verification where the data collected with CoTEVer can requires additional knowledge which makes it be utilized for enhancing the faithfulness of hard for annotators to intuitively write a revised explanation explanations. Our toolkit is publicly available and answer.
arXiv.org Artificial Intelligence
Mar-6-2023
- Country:
- North America > United States (0.69)
- Genre:
- Research Report (0.40)
- Workflow (0.46)
- Industry:
- Automobiles & Trucks > Manufacturer (0.30)
- Health & Medicine (0.68)
- Technology: