TheUnreliabilityofExplanationsinFew-shot PromptingforTextualReasoning
–Neural Information Processing Systems
However, text-davinci-002 is able to benefit more substantially. We further show that explanations generated by the LLMs may not entail the models' predictions norbefactually grounded intheinput, evenonsimple tasks with extractive explanations. However, these flawed explanations can still be useful as a way to verify LLMs' predictions post-hoc.
Neural Information Processing Systems
Feb-11-2026, 19:26:33 GMT