What is RLHF? In some sense, RLHF is part of that alignment process, where you tune the model to look at a bunch of human preferences.
You can show a bunch of example outputs and then let users, actual humans, decide what output is better for them, and you tune the model.
RLHF: Aligning AI Models with Human Preferences
By
–
Leave a Reply