Llama 3.2 is here and we're faster than ever! We've been independently verified as the fastest #AI #Inference on @AIatMeta
's Llama 3.2 1B & 3B with … 2470 tokens/sec on 1B 1566 tokens/sec on 3B … all running at full-precision! Start developing
Llama 3.2 Launches with Record-Breaking Inference Speed Performance
By
–
Leave a Reply