PRESERVE: Prefetching Model Weights and KV-Cache in Distributed LLM Serving