MNN-LLM: A Generic Inference Engine for Fast Large Language Model Deployment on Mobile Devices