QuickLLaMA: Query-aware Inference Acceleration for Large Language Models