One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments