Bridging the Knowledge-Prediction Gap in LLMs on Multiple-Choice Questions
Park, Yoonah, Pyun, Haesung, Jo, Yohan
–arXiv.org Artificial Intelligence
Large Language Models (LLMs) often fail on multiple-choice questions (MCQs) despite demonstrating correct knowledge in other contexts, such as free-form generation. To investigate the mechanism underlying this knowledge-prediction gap on MCQs and alleviate it, we conduct a probing analysis and find that residual streams in certain layers contain a subspace spanned by two important bases: a \emph{knowledge basis} that encodes the probability of the ground-truth answer for a given MCQ and a \emph{prediction basis} that encodes the probability of the answer choice predicted by the model. We observe that incorrect predictions arise from a misalignment of the model's hidden states along these two bases. Hence, we introduce \textbf{KAPPA} (Knowledge-Aligned Prediction through Projection-based Adjustment), a parameter-free intervention that transforms the hidden states to align the prediction coordinate with the knowledge coordinate within this subspace. Experiments on binary-choice reformulations of Big-Bench-Hard and ARC-Challenge show that KAPPA substantially improves accuracy and consistently outperforms baselines. While optimal subspaces differ across tasks, subspaces generalize to some extent, as supported by cross-dataset experiments. Moreover, KAPPA extends its effectiveness to free-form questions beyond MCQs. Our work provides a new geometric understanding of the knowledge-prediction gap and offers a practical method for better aligning model behavior with its latent knowledge.
arXiv.org Artificial Intelligence
Dec-10-2025
- Country:
- Asia (0.93)
- Europe > Austria (0.28)
- North America
- Canada (0.28)
- Mexico (0.28)
- United States (0.46)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Education > Curriculum > Subject-Specific Education (1.00)
- Technology: