Compressing Chain-of-Thought in LLMs via Step Entropy