AI Dynamics

Global AI News Aggregator

FT-GPT-3.5 Latency Reduced 4-5x Faster Inference Speed

We’ve reduced the model latency by 4-5x, serving results on average in 0.65 seconds instead of 3.15 seconds (FT-GPT-3.5 compared to GPT-4). You may notice the speedup when Copilot prompts you for user input. Every second counts, and we’re here to make them all productive.

→ View original post on X — @perplexity_ai,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *