From Prompts to Power: Measuring the Energy Footprint of LLM Inference