SmoothRot: Combining Channel-Wise Scaling and Rotation for Quantization-Friendly LLMs