Thinking, Faithful and Stable: Mitigating Hallucinations in LLMs