Approximating Human Preferences Using a Multi-Judge Learned System
Sprejer, Eitán, Avalos, Fernando, Bernardi, Augusto, Faustino, Jose Pedro Brito de Azevedo, Haimes, Jacob, Oozeer, Narmeen Fatimah
–arXiv.org Artificial Intelligence
Aligning LLM-based judges with human preferences is a significant challenge, as they are difficult to calibrate and often suffer from rubric sensitivity, bias, and instability. Overcoming this challenge advances key applications, such as creating reliable reward models for Reinforcement Learning from Human Feedback (RLHF) and building effective routing systems that select the best-suited model for a given user query. In this work, we propose a framework for modeling diverse, persona-based preferences by learning to aggregate outputs from multiple rubric-conditioned judges. We investigate the performance of this approach against naive baselines and assess its robustness through case studies on both human and LLM-judges biases. Our primary contributions include a persona-based method for synthesizing preference labels at scale and two distinct implementations of our aggregator: Generalized Additive Model (GAM) and a Multi-Layer Perceptron (MLP).
arXiv.org Artificial Intelligence
Oct-31-2025
- Country:
- Asia > Myanmar
- Tanintharyi Region > Dawei (0.04)
- Europe > Czechia
- Pardubice Region > Pardubice (0.04)
- North America > United States
- California > San Francisco County
- San Francisco (0.14)
- Colorado > Boulder County
- Boulder (0.04)
- California > San Francisco County
- South America
- Argentina > Pampas
- Buenos Aires F.D. > Buenos Aires (0.04)
- Brazil > São Paulo (0.04)
- Colombia > Bogotá D.C.
- Bogotá (0.04)
- Argentina > Pampas
- Asia > Myanmar
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Research Report
- Industry:
- Information Technology > Security & Privacy (1.00)
- Law (0.93)
- Technology: