Transforming and Combining Rewards for Aligning Large Language Models