EAC-MoE: Expert-Selection Aware Compressor for Mixture-of-Experts Large Language Models