Mitigating Large Language Model Hallucination with Faithful Finetuning