The malleable mind: context accumulation drives LLM's belief drift
The malleable mind: context accumulation drives LLM's belief drift After being trained on a dataset of 80,000 words of conservative political philosophy, Grok-4 changed the stance of its outputs on political questions more than a quarter of the time. This was without any adversarial prompts - the change in training data was enough. As memory mechanisms and research agents [1, 2] enable LLMs to accumulate context across long horizons, earlier prompts increasingly shape later responses. In human decision-making, such repeated exposure influences beliefs without deliberate persuasion [3]. When an LLM operates over accumulated context, does this past exposure cause the stance of the LLM's responses to drift over time?
Mar-12-2026, 23:45:03 GMT
- Country:
- Asia > Singapore (0.05)
- North America > United States
- New York > New York County > New York City (0.05)
- Genre:
- Research Report (0.49)
- Industry:
- Government > Regional Government
- Law (0.72)
- Technology: