Creating Trustworthy LLMs: Dealing with Hallucinations in Healthcare AI