Dynamic Expert Quantization for Scalable Mixture-of-Experts Inference
Chu, Kexin, Xiang, Dawei, Shen, Zixu, Yang, Yiwei, Liu, Zecheng, Zhang, Wei
–arXiv.org Artificial Intelligence
Mixture-of-Experts (MoE) models scale LLM capacity efficiently, but deployment on consumer GPUs is limited by the large memory footprint of inactive experts. Static post-training quantization reduces storage costs but cannot adapt to shifting activation patterns, causing accuracy loss under aggressive compression. So we present DynaExq, a runtime system that treats expert precision as a first-class, dynamically managed resource. DynaExq combines (1) a hotness-aware precision controller that continuously aligns expert bit-widths with long-term activation statistics, (2) a fully asynchronous precision-switching pipeline that overlaps promotion and demotion with MoE computation, and (3) a fragmentation-free memory pooling mechanism that supports hybrid-precision experts with deterministic allocation. Together, these components enable stable, non-blocking precision transitions under strict HBM budgets. Across Qwen3-30B and Qwen3-80B MoE models and six representative benchmarks, DynaExq deploys large LLMs on single RTX 5090 and A6000 GPUs and improves accuracy by up to 4.03 points over static low-precision baselines. The results show that adaptive, workload-aware quantization is an effective strategy for memory-constrained MoE serving.
arXiv.org Artificial Intelligence
Nov-25-2025
- Country:
- Asia
- China (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.41)
- North America > United States
- California > Santa Cruz County
- Santa Cruz (0.14)
- Connecticut > Tolland County
- Storrs (1.00)
- California > Santa Cruz County
- Asia
- Genre:
- Research Report > New Finding (0.34)
- Technology: