Is DeepSeek V3.1 a game-changer or mostly hype? Let’s break it down 👇
— Louis-François Bouchard 🎥🤖 (@Whats_AI) 21 août 2025
What’s new:
• Still a similar MoE transformer (671B params, 37B active) with 128K context
• Merges DeepSeek V3 + R1 → one hybrid model with thinking (reasoning) and non-thinking (direct) modes — like… pic.twitter.com/WKMPGwGlQM
Is DeepSeek V3.1 a game-changer or mostly hype? Let’s break it down What’s new: • Still a similar MoE transformer (671B params, 37B active) with 128K context • Merges DeepSeek V3 + R1 → one hybrid model with thinking (reasoning) and non-thinking (direct) modes — like
Leave a Reply