Reasoning-to-Defend: Safety-Aware Reasoning Can Defend Large Language Models from Jailbreaking