Reinforcement Learning from Human Feedback: Whose Culture, Whose Values, Whose Perspectives?
Barman, Kristian González, Lohse, Simon, de Regt, Henk
–arXiv.org Artificial Intelligence
This approach is partic ularly useful when designing AI systems for tasks where it is difficult to specify a precise reward function or when it is important to align the model's behaviour with certain human expectations and values. For instance, RLHF has notably improved language models for context - aware text generation (Ziegler et al. 2020) and taught robots to navigate cluttered environments (Henry et al. 2010) . RLHF is commonly employed in the later stages of fine - tuning models, particularly in the development of prominent Large Language Models (LLMs) like GPT - 3.5 or GPT - 4. Initially, these models undergo training using vast text corpora to grasp a broad range of language patterns and contexts. This foundational training is supplemented by task - specific fine - tuning, where the models are adjusted to excel in particular applications, such as understanding and generating dialogues. The refinement process is then furt her enhanced through RLHF.
arXiv.org Artificial Intelligence
Jul-2-2024
- Country:
- Africa > Middle East (0.04)
- Asia > Middle East (0.04)
- Europe
- Belgium > Flanders
- East Flanders > Ghent (0.04)
- Middle East (0.04)
- Netherlands (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- Belgium > Flanders
- North America
- Canada (0.04)
- United States
- California > Santa Clara County
- Palo Alto (0.04)
- Illinois > Cook County
- Chicago (0.04)
- New Jersey > Mercer County
- Princeton (0.04)
- New York > New York County
- New York City (0.04)
- California > Santa Clara County
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine > Epidemiology (0.46)
- Information Technology > Security & Privacy (0.46)
- Law > Civil Rights & Constitutional Law (0.67)
- Leisure & Entertainment > Games (0.46)
- Technology: