Efficient Reinforcement Learning from Human Feedback via Bayesian Preference Inference
Cercola, Matteo, Capretti, Valeria, Formentin, Simone
–arXiv.org Artificial Intelligence
Learning from human preferences is a cornerstone of aligning machine learning models with subjective human judgments. Yet, collecting such preference data is often costly and time-consuming, motivating the need for more efficient learning paradigms. Two established approaches offer complementary advantages: RLHF scales effectively to high-dimensional tasks such as LLM fine-tuning, while PBO achieves greater sample efficiency through active querying. We propose a hybrid framework that unifies RLHF's scalability with PBO's query efficiency by integrating an acquisition-driven module into the RLHF pipeline, thereby enabling active and sample-efficient preference gathering. We validate the proposed approach on two representative domains: (i) high-dimensional preference optimization and (ii) LLM fine-tuning. Experimental results demonstrate consistent improvements in both sample efficiency and overall performance across these tasks.
arXiv.org Artificial Intelligence
Nov-7-2025
- Country:
- Europe > Italy
- North America
- Canada > British Columbia (0.04)
- United States > Pennsylvania
- Allegheny County > Pittsburgh (0.04)
- Genre:
- Research Report (0.70)
- Technology: