MaxMin-RLHF: Towards Equitable Alignment of Large Language Models with Diverse Human Preferences

Open in new window