Emergence of Episodic Memory in Transformers: Characterizing Changes in Temporal Structure of Attention Scores During Training
Mistry, Deven Mahesh, Bajaj, Anooshka, Aggarwal, Yash, Maini, Sahaj Singh, Tiganj, Zoran
–arXiv.org Artificial Intelligence
We investigate in-context temporal biases in attention heads and transformer outputs. Using cognitive science methodologies, we analyze attention scores and outputs of the GPT-2 models of varying sizes. Across attention heads, we observe effects characteristic of human episodic memory, including temporal contiguity, primacy and recency. Transformer outputs demonstrate a tendency toward in-context serial recall. Importantly, this effect is eliminated after the ablation of the induction heads, which are the driving force behind the contiguity effect. Our findings offer insights into how transformers organize information temporally during in-context learning, shedding light on their similarities and differences with human memory and learning.
arXiv.org Artificial Intelligence
Feb-9-2025
- Country:
- Asia > Middle East
- UAE (0.14)
- North America > United States (0.28)
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Health & Medicine > Consumer Health (0.61)
- Technology: