Zero-Delay QKV Compression for Mitigating KV Cache and Network Bottlenecks in LLM Inference