KeepKV: Achieving Periodic Lossless KV Cache Compression for Efficient LLM Inference