THOUGHTSCULPT: Reasoning with Intermediate Revision and Search
Chi, Yizhou, Yang, Kevin, Klein, Dan
–arXiv.org Artificial Intelligence
Whilst Large Language Models (LLMs) such as GPT (Brown et al., 2020; OpenAI, 2024), LLaMA (Touvron et al., 2023a;b), and Claude (Anthropic, 2024) are developed to be increasingly capable in performing a variety of reasoning tasks, recent studies have revealed that the utilization of distinct prompting strategies and instructional guidance can have a notable influence on the performance of LLMs when tackling identical tasks. Chain-of-Thought (CoT) is a prompting strategy detailed in (Wei et al., 2023) that directs LLMs to produce the final task output through intermediate steps of reasoning, referred to as "intermediate thoughts." Notably, CoT has demonstrated a substantial enhancement in the problem-solving proficiency of LLMs without necessitating any model updates. Self-consistency with CoT (CoT-SC) (Wang et al., 2023a) is proposed to improve output consistency by generating multiple CoTs and selecting the best outcome. Recently, in extension to CoT and CoT-SC, Tree-of-Thoughts (Yao et al., 2023a) and Graph-of-Thoughts (Besta et al., 2024) are proposed to shape the reasoning process of LLMs as a tree or an arbitrary graph structure. These approaches enable LLMs to explore different paths of thought and find better outputs by utilizing backtracking and graph-search algorithms. However, these approaches' reasoning capabilities are often limited by the set of candidates they generate at earlier steps. They cannot revise and edit their original answers continuously in later steps. As a result, these methods may not be effective in addressing problems that require frequent revision and modifications.
arXiv.org Artificial Intelligence
Apr-8-2024
- Country:
- North America > United States (0.93)
- Genre:
- Research Report (1.00)
- Industry:
- Leisure & Entertainment (1.00)
- Technology: