Mustafar: Promoting Unstructured Sparsity for KV Cache Pruning in LLM Inference