LMCache: An Efficient KV Cache Layer for Enterprise-Scale LLM Inference