Unlocking Efficient Long-to-Short LLM Reasoning with Model Merging