$A^3$: Attention-Aware Accurate KV Cache Fusion for Fast Large Language Model Serving