AI Dynamics

Global AI News Aggregator

Gemma’s 256K Token Context: Analyzing Language Distribution Patterns

Keep in mind Gemma has a 256K tokens, so all text could be quite a bit shorter because there are so many merges. The interesting analysis here is to look at the *distribution* of token counts across different languages, and compare it to that same distribution for previous.

→ View original post on X — @karpathy,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *