Events & Conferences2 months ago
A better training method for reinforcement learning with human feedback
Reinforcement learning with human feedback (RLHF) is the standard method for aligning large language models (LLMs) with human preferences — such as the preferences for nontoxic...