From @eetimes
, @SallyWardFoxton covered our cloud inference offering stating: "For the larger Llama3.1-70B, SambaNova is currently claiming the crown with 580 tokens/s versus Cerebras’ 445 and Groq’s 544 tokens/s.” Our CEO @RodrigoLiang
, was also quoted “As we [the industry]
SambaNova Leads Cloud Inference Performance with Llama 3.1-70B
By
–
Leave a Reply