Expressive and Scalable Quantum Fusion for Multimodal Learning
Nguyen, Tuyen, Hoang, Trong Nghia, Nguyen, Phi Le, Vu, Hai L., Thang, Truong Cong
–arXiv.org Artificial Intelligence
The aim of this paper is to introduce a quantum fusion mechanism for multimodal learning and to establish its theoretical and empirical potential. The proposed method, called the Quantum Fusion Layer (QFL), replaces classical fusion schemes with a hybrid quantum-classical procedure that uses parameterized quantum circuits to learn entangled feature interactions without requiring exponential parameter growth. Supported by quantum signal processing principles, the quantum component efficiently represents high-order polynomial interactions across modalities with linear parameter scaling, and we provide a separation example between QFL and low-rank tensor-based methods that highlights potential quantum query advantages. In simulation, QFL consistently outperforms strong classical baselines on small but diverse multimodal tasks, with particularly marked improvements in high-modality regimes. These results suggest that QFL offers a fundamentally new and scalable approach to multimodal fusion that merits deeper exploration on larger systems.
arXiv.org Artificial Intelligence
Oct-9-2025
- Country:
- Asia
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America > United States
- California > Los Angeles County
- Los Angeles (0.04)
- Washington (0.04)
- California > Los Angeles County
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Genre:
- Research Report > New Finding (0.87)
- Industry:
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning > Neural Networks
- Deep Learning (0.93)
- Natural Language (1.00)
- Representation & Reasoning > Information Fusion (0.68)
- Robots (0.67)
- Vision (1.00)
- Machine Learning > Neural Networks
- Communications (0.93)
- Data Science (1.00)
- Hardware (1.00)
- Artificial Intelligence
- Information Technology