Corrected Calculations (the original claim doesn't change, as it goes from 5e+24 -> 1.1e+24 and is still two orders of magnitude behind):
LLaMa2-70B was trained on 2T tokens, and 1.7m hours of A100 GPU time. At a HFU of ~60%, LLama2-70B took ~1.1e+24 flops (1.7m * 312TFlops
LLaMa2-70B Training Computation Corrected Calculations
By
–
Leave a Reply