Amber Pruner: Leveraging N:M Activation Sparsity for Efficient Prefill in Large Language Models