Mitigating KV Cache Competition to Enhance User Experience in LLM Inference