SlimInfer: Accelerating Long-Context LLM Inference via Dynamic Token Pruning