Self-Consistency Improves Chain of Thought Reasoning in Language Models
Wang, Xuezhi, Wei, Jason, Schuurmans, Dale, Le, Quoc, Chi, Ed, Narang, Sharan, Chowdhery, Aakanksha, Zhou, Denny
–arXiv.org Artificial Intelligence
Chain-of-thought prompting combined with pre-trained large language models has achieved encouraging results on complex reasoning tasks. In this paper, we propose a new decoding strategy, self-consistency, to replace the naive greedy decoding used in chain-of-thought prompting. It first samples a diverse set of reasoning paths instead of only taking the greedy one, and then selects the most consistent answer by marginalizing out the sampled reasoning paths. Self-consistency leverages the intuition that a complex reasoning problem typically admits multiple different ways of thinking leading to its unique correct answer. Our extensive empirical evaluation shows that self-consistency boosts the performance of chain-of-thought prompting with a striking margin on a range of popular arithmetic and commonsense reasoning benchmarks, including GSM8K (+17.9%), Although language models have demonstrated remarkable success across a range of NLP tasks, their ability to demonstrate reasoning is ...
arXiv.org Artificial Intelligence
Mar-7-2023
- Country:
- Asia > Middle East (0.67)
- Europe (1.00)
- North America > United States (0.67)
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine > Therapeutic Area > Oncology (0.68)
- Technology: