LoRAP: Transformer Sub-Layers Deserve Differentiated Structured Compression for Large Language Models