MNN-AECS: Energy Optimization for LLM Decoding on Mobile Devices via Adaptive Core Selection
Huang, Zhengxiang, Niu, Chaoyue, Wang, Zhaode, Xue, Jiarui, Zhang, Hanming, Wang, Yugang, Xin, Zewei, Jiang, Xiaotang, Lv, Chengfei, Wu, Fan, Chen, Guihai
–arXiv.org Artificial Intelligence
As the demand for on-device Large Language Model (LLM) inference grows, energy efficiency has become a major concern, especially for battery-limited mobile devices. Our analysis shows that the memory-bound LLM decode phase dominates energy use, and yet most existing works focus on accelerating the prefill phase, neglecting energy concerns. We introduce Adaptive Energy-Centric Core Selection (AECS) and integrate it into MNN to create the energy-efficient version, MNN-AECS, the first engine-level system solution without requiring root access or OS modifications for energy-efficient LLM decoding. MNN-AECS is designed to reduce LLM decoding energy while keeping decode speed within an acceptable slowdown threshold by dynamically selecting low-power CPU cores. MNN-AECS is evaluated across 5 Android and 2 iOS devices on 5 popular LLMs of various sizes. Compared to original MNN, MNN-AECS cuts down energy use by 23% without slowdown averaged over all 7 devices and 4 datasets. Against other engines, including llama.cpp, executorch, mllm, and MediaPipe, MNN-AECS delivers 39% to 78% energy saving and 12% to 363% speedup on average.
arXiv.org Artificial Intelligence
Jun-26-2025
- Country:
- Africa > Mali (0.04)
- Asia
- China
- Shanghai > Shanghai (0.05)
- Zhejiang Province > Hangzhou (0.04)
- Middle East > Jordan (0.04)
- China
- Europe
- Germany (0.04)
- Netherlands > South Holland
- Rotterdam (0.04)
- Spain > Galicia
- Madrid (0.04)
- North America > United States
- California
- District of Columbia > Washington (0.05)
- Hawaii > Honolulu County
- Honolulu (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- New York > New York County
- New York City (0.05)
- Genre:
- Research Report (0.64)
- Industry:
- Energy (1.00)
- Technology: