Do Language Models Know When They're Hallucinating References?