TinyServe: Query-Aware Cache Selection for Efficient LLM Serving