Unfamiliar Finetuning Examples Control How Language Models Hallucinate