EDiT: Efficient Diffusion Transformers with Linear Compressed Attention