Shadow in the Cache: Unveiling and Mitigating Privacy Risks of KV-cache in LLM Inference