Leaner Training, Lower Leakage: Revisiting Memorization in LLM Fine-Tuning with LoRA