Accelerating LLM Inference Throughput via Asynchronous KV Cache Prefetching