KV-Runahead: Scalable Causal LLM Inference by Parallel Key-Value Cache Generation