When Parts are Greater Than Sums: Individual LLM Components Can Outperform Full Models
Chang, Ting-Yun, Thomason, Jesse, Jia, Robin
–arXiv.org Artificial Intelligence
This paper studies in-context learning (ICL) by decomposing the output of large language models into the individual contributions of attention heads and MLPs (components). We observe curious components: good-performing ones that individually do well on a classification task, even when the model performs poorly; bad-performing ones that do much worse than chance; and label-biased components that always predict the same label. We find that component accuracies are well-correlated across different demonstration sets and perturbations of prompt templates, even when the full-model accuracy varies greatly. Based on our findings, we propose component reweighting, which learns to linearly re-scale the component activations from a few labeled examples. Given 24 labeled examples, our method improves by an average of 6.0% accuracy points over 24-shot ICL across 8 tasks on Llama-2-7B. Overall, this paper both enriches our understanding of ICL and provides a practical method for improvement Figure 1: Each dot represents a component (attention by examining model internals.
arXiv.org Artificial Intelligence
Jun-24-2024
- Country:
- Asia (0.68)
- Europe (0.68)
- North America > United States
- California > Los Angeles County
- Los Angeles (0.14)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Washington > King County
- Seattle (0.14)
- California > Los Angeles County
- Genre:
- Research Report
- Experimental Study (0.46)
- New Finding (0.48)
- Research Report
- Technology: