Rewarding Doubt: A Reinforcement Learning Approach to Confidence Calibration of Large Language Models