Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language Models