ModernBERT ftw! @answerdotai & @LightOnIO killing it!! > ModernBERT-base: 22 layers, 149M params
> ModernBERT-large: 28 layers, 395M params > 2 trillion tokens of English and code data.
> Up to 8,192 tokens, ideal for processing long documents
> RoPE for long-context support
ModernBERT: New Open-Source Language Model from Answer.ai
By
–
Leave a Reply