Beyond Token Length: Step Pruner for Efficient and Accurate Reasoning in Large Language Models