AI Dynamics

Global AI News Aggregator

Pre-training and Instruction-Tuning in Large Language Models

The pre-training stage is where the billions of words of training data come into play – the instruction-tuning / RLHF stage is where human labelers are asked to vote on which generations are "best" – that's the bit that might influence things like "delve" https://
openai.com/research/instr
uction-following

→ View original post on X — @simonw,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *