Head-Tail-Aware KL Divergence in Knowledge Distillation for Spiking Neural Networks