EG-MLA: Embedding-Gated Multi-head Latent Attention for Scalable and Efficient LLMs