Hallucination Detection: Robustly Discerning Reliable Answers in Large Language Models