Robust Reinforcement Learning from Corrupted Human Feedback Alexander Bukharin

Neural Information Processing Systems 

Reinforcement learning from human feedback (RLHF) provides a principled framework for aligning AI systems with human preference data.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found