The Alignment Ceiling: Objective Mismatch in Reinforcement Learning from Human Feedback