"The lesson is post-training bottlenecks are solved by refining methods and data" Zhibin Gou (@zebgou) If Gemini-3 proved continual scaling pretraining, DeepSeek-V3.2-Speciale proves scaling RL with large context. We spent a year pushing DeepSeek-V3 to its limits. The lesson is post-training bottlenecks are solved by refining methods and data, not just waiting for a better base. — https://nitter.net/zebgou/status/1995462720078934213#m
Post-training bottlenecks solved through refined methods and data
By
–
Leave a Reply