EvolKV: Evolutionary KV Cache Compression for LLM Inference