AI Dynamics

Global AI News Aggregator

Meta’s Continuous Chain-of-Thought Reasoning for LLMs

Can reasoning LLMs think better if their Chain-of-Thought is continuous instead of discrete? This Meta paper introduces the first scalable way to train continuous CoTs with reinforcement learning—no need to distill from discrete references. By using "soft" tokens

→ View original post on X — @jiqizhixin,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *