Rule-Guided Feedback: Enhancing Reasoning by Enforcing Rule Adherence in Large Language Models
Diallo, Aissatou, Bikakis, Antonis, Dickens, Luke, Hunter, Anthony, Miller, Rob
–arXiv.org Artificial Intelligence
In this paper, we introduce Rule-Guided Feedback (RGF), a framework designed to enhance Large Language Model (LLM) performance through structured rule adherence and strategic information seeking. RGF implements a teacher-student paradigm where rule-following is forced through established guidelines. Our framework employs a Teacher model that rigorously evaluates each student output against task-specific rules, providing constructive guidance rather than direct answers when detecting deviations. This iterative feedback loop serves two crucial purposes: maintaining solutions within defined constraints and encouraging proactive information seeking to resolve uncertainties. We evaluate RGF on diverse tasks including Checkmate-in-One puzzles, Sonnet Writing, Penguins-In-a-Table classification, GSM8k, and StrategyQA. Our findings suggest that structured feedback mechanisms can significantly enhance LLMs' performance across various domains.
arXiv.org Artificial Intelligence
Mar-14-2025
- Country:
- Europe > United Kingdom (0.14)
- Genre:
- Research Report > New Finding (0.86)
- Industry:
- Health & Medicine (0.46)
- Leisure & Entertainment (0.47)
- Technology: