We also briefly tried instruction finetuning using the approach of Chung et al. (2022).
The resulting model, LLaMA-I, outperforms Flan-PaLM-cont (62B) on MMLU and showcases some interesting instruct capabilities.
7/n
Instruction Finetuning Results: LLaMA-I Outperforms Flan-PaLM
By
–
Leave a Reply