KVSink: Understanding and Enhancing the Preservation of Attention Sinks in KV Cache Quantization for LLMs