Accelerating LLM Inference with Precomputed Query Storage