Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs