LLM Cache Bandit Revisited: Addressing Query Heterogeneity for Cost-Effective LLM Inference
Yang, Hantao, Xie, Hong, Lian, Defu, Chen, Enhong
–arXiv.org Artificial Intelligence
This paper revisits the LLM cache bandit problem, with a special focus on addressing the query heterogeneity for cost-effective LLM inference. Previous works often assume uniform query sizes. Heterogeneous query sizes introduce a combinatorial structure for cache selection, making the cache replacement process more computationally and statistically challenging. We treat optimal cache selection as a knapsack problem and employ an accumulation-based strategy to effectively balance computational overhead and cache updates. In theoretical analysis, we prove that the regret of our algorithm achieves an $O(\sqrt{MNT})$ bound, improving the coefficient of $\sqrt{MN}$ compared to the $O(MN\sqrt{T})$ result in Berkeley, where $N$ is the total number of queries and $M$ is the cache size. Additionally, we also provide a problem-dependent bound, which was absent in previous works. The experiment rely on real-world data show that our algorithm reduces the total cost by approximately 12\%.
arXiv.org Artificial Intelligence
Sep-22-2025
- Country:
- Asia
- China (0.04)
- Japan > Honshū
- Tōhoku > Iwate Prefecture > Morioka (0.04)
- Middle East > Jordan (0.04)
- Singapore (0.04)
- North America > United States (0.04)
- South America > Chile
- Asia
- Genre:
- Research Report (0.82)
- Technology: