MossNet: Mixture of State-Space Experts is a Multi-Head Attention
Tuli, Shikhar, Smith, James Seale, Jeelani, Haris, Lin, Chi-Heng, Patel, Abhishek, Ramanishka, Vasili, Hsu, Yen-Chang, Jin, Hongxia
–arXiv.org Artificial Intelligence
Large language models (LLMs) have significantly advanced generative applications in natural language processing (NLP). Recent trends in model architectures revolve around efficient variants of transformers or state-space/gated-recurrent models (SSMs, GRMs). However, prevailing SSM/GRM-based methods often emulate only a single attention head, potentially limiting their expressiveness. In this work, we propose MossNet, a novel mixture-of-state-space-experts architecture that emulates a linear multi-head attention (MHA). MossNet leverages a mixture-of-experts (MoE) implementation not only in channel-mixing multi-layered perceptron (MLP) blocks but also in the time-mixing SSM kernels to realize multiple "attention heads." Extensive experiments on language modeling and downstream evaluations show that MossNet outperforms both transformer- and SSM-based architectures of similar model size and data budgets. Larger variants of MossNet, trained on trillions of tokens, further confirm its scalability and superior performance. In addition, real-device profiling on a Samsung Galaxy S24 Ultra and an Nvidia A100 GPU demonstrate favorable runtime speed and resource usage compared to similarly sized baselines. Our results suggest that MossNet is a compelling new direction for efficient, high-performing recurrent LLM architectures.
arXiv.org Artificial Intelligence
Oct-31-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > United States
- California > Santa Clara County > Mountain View (0.04)
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.54)
- Industry:
- Information Technology (0.54)
- Technology: