PEA: Enhancing LLM Performance on Computational-Reasoning Tasks
Wang, Zi, Weng, Shiwei, Alhanahnah, Mohannad, Jha, Somesh, Reps, Tom
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) have exhibited significant generalization capabilities across diverse domains, prompting investigations into their potential as generic reasoning engines. Recent studies have explored inference-time computation techniques [Welleck et al., 2024, Snell et al., 2024], particularly prompt engineering methods such as Chain-of-Thought (CoT), to enhance LLM performance on complex reasoning tasks [Wei et al., 2022]. These approaches have successfully improved model performance and expanded LLMs' practical applications. However, despite the growing focus on enhancing model capabilities through inference-time computation for complex reasoning tasks, the current literature lacks a formal framework to precisely describe and characterize the complexity of reasoning problems. This study identifies a class of reasoning problems, termed computational reasoning problems, which are particularly challenging for LLMs [Yao et al., 2023, Hao et al., 2024, Valmeekam et al., 2023], such as planning problems and arithmetic games. Informally, these problems can be accurately described using succinct programmatic representations. We propose a formal framework to describe and algorithmically solve these problems. The framework employs first-order logic, equipped with efficiently computable predicates and finite domains.
arXiv.org Artificial Intelligence
Feb-15-2025
- Country:
- North America > United States (0.93)
- Genre:
- Research Report (1.00)
- Industry:
- Transportation > Air (0.34)
- Technology: