With GPT-3, didn't have a way to steer the model's behavior. Now we pick a clusters of issues (e.g. model doubling down on wrong answers — something that used to be a huge problem), build evals, collect data, retrain the model weekly, and deploy. Anyone can test if we've improved
Steering GPT behavior through weekly model retraining and evaluation
By
–
Leave a Reply