Efficient Long-Context LLM Inference via KV Cache Clustering