From Proof to Program: Characterizing Tool-Induced Reasoning Hallucinations in Large Language Models