Memory Augmented Large Language Models are Computationally Universal
–arXiv.org Artificial Intelligence
We show that transformer-based large language models are computationally universal when augmented with an external memory. Any deterministic language model that conditions on strings of bounded length is equivalent to a finite automaton, hence computationally limited. However, augmenting such models with a read-write memory creates the possibility of processing arbitrarily large inputs and, potentially, simulating any algorithm. We establish that an existing large language model, Flan-U-PaLM 540B, can be combined with an associative read-write memory to exactly simulate the execution of a universal Turing machine, $U_{15,2}$. A key aspect of the finding is that it does not require any modification of the language model weights. Instead, the construction relies solely on designing a form of stored instruction computer that can subsequently be programmed with a specific set of prompts.
arXiv.org Artificial Intelligence
Jan-9-2023
- Country:
- Europe
- Ireland (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- North America > Canada
- Alberta (0.14)
- Europe
- Genre:
- Research Report (0.50)
- Technology: