Quantum Large Language Models via Tensor Network Disentanglers
Aizpurua, Borja, Jahromi, Saeed S., Singh, Sukhbinder, Orus, Roman
–arXiv.org Artificial Intelligence
We propose a method to enhance the performance of Large Language Models (LLMs) by integrating quantum computing and quantum-inspired techniques. Specifically, our approach involves replacing the weight matrices in the Self-Attention and Multi-layer Perceptron layers with a combination of two variational quantum circuits and a quantum-inspired tensor network, such as a Matrix Product Operator (MPO). This substitution enables the reproduction of classical LLM functionality by decomposing weight matrices through the application of tensor network disentanglers and MPOs, leveraging well-established tensor network techniques. By incorporating more complex and deeper quantum circuits, along with increasing the bond dimensions of the MPOs, our method captures additional correlations within the quantum-enhanced LLM, leading to improved accuracy beyond classical models while maintaining low memory overhead.
arXiv.org Artificial Intelligence
Oct-22-2024
- Country:
- Africa > Central African Republic
- Ombella-M'Poko > Bimbo (0.04)
- Europe > Spain
- Basque Country > Biscay Province > Bilbao (0.04)
- North America > Canada
- Africa > Central African Republic
- Genre:
- Research Report (0.64)
- Technology: