The Few-shot Dilemma: Over-prompting Large Language Models
Tang, Yongjian, Tuncel, Doruk, Koerner, Christian, Runkler, Thomas
–arXiv.org Artificial Intelligence
Abstract--Over-prompting, a phenomenon where excessive examples in prompts lead to diminished performance in Large Language Models (LLMs), challenges the conventional wisdom about in-context few-shot learning. T o investigate this few-shot dilemma, we outline a prompting framework that leverages three standard few-shot selection methods - random sampling, semantic embedding, and TF-IDF vectors - and evaluate these methods across multiple LLMs, including GPT -4o, GPT -3.5-turbo, DeepSeek-V3, Gemma-3, LLaMA-3.1, Our experimental results reveal that incorporating excessive domain-specific examples into prompts can paradoxically degrade performance in certain LLMs, which contradicts the prior empirical conclusion that more relevant few-shot examples universally benefit LLMs. Given the trend of LLM-assisted software engineering and requirement analysis, we experiment with two real-world software requirement classification datasets. By gradually increasing the number of TF-IDF-selected and stratified few-shot examples, we identify their optimal quantity for each LLM. This combined approach achieves superior performance with fewer examples, avoiding the over-prompting problem, thus surpassing the state-of-the-art by 1% in classifying functional and non-functional requirements. Instruction-tuned LLMs have demonstrated exceptional language understanding and knowledge inference capabilities [12].
arXiv.org Artificial Intelligence
Sep-17-2025
- Country:
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Genre:
- Research Report (1.00)
- Technology: