Simultaneous Multi-objective Alignment Across Verifiable and Non-verifiable Rewards
Shen, Yiran, Xia, Yu, Chang, Jonathan, Ammanabrolu, Prithviraj
–arXiv.org Artificial Intelligence
Aligning large language models to human preferences is inherently multidimensional, yet most pipelines collapse heterogeneous signals into a single optimizeable objective. We seek to answer what it would take to simultaneously align a model across various domains spanning those with: verifiable rewards (mathematical accuracy), non-verifiable subjective preferences (human values), and complex interactive scenarios (multi-turn AI tutoring dialogues). Such multi-objective reinforcement learning setups are often plagued by the individual objectives being at odds with each other, resulting in inefficient training and little user control during inference. We propose a unified framework that: (i) standardizes {process reward model} (PRM) training across both verifiable and non-verifiable settings to better supervise models' chain-of-thought reasoning; (ii) performs {multi-objective alignment} by training the LLM with our $\textbf{M}$ulti-$\textbf{A}$ction-$\textbf{H}$ead $\textbf{DPO}$ (MAH-DPO) and a vectorized reward where the dimensions of the vector correspond to the various objectives instead of a single scalar; and (iii) demonstrates how such a system provides fine-grained inference-time user control. Experiments across math reasoning, value alignment, and multi-turn dialogue show that our framework improves performance across multiple objectives simultaneously, while minimizing cross-objective trade-offs and enabling flexible inference time user control. The code can be found at https://github.com/pearls-lab/multiobj-align.
arXiv.org Artificial Intelligence
Oct-2-2025
- Country:
- North America > United States
- California > San Diego County
- San Diego (0.04)
- Virginia (0.04)
- California > San Diego County
- North America > United States
- Genre:
- Research Report > New Finding (0.46)
- Workflow (0.67)
- Industry:
- Education (1.00)
- Technology: