LLMs Know What to Drop: Self-Attention Guided KV Cache Eviction for Efficient Long-Context Inference