It's never too late for a cheeky Kimi K2.5 fine-tune
@maximelabonne
-

Spicy Takes on Small Language Models at AI Engineer
By
–
See you on Thursday at @aiDotEngineer for some spicy takes on small language models! π«‘ I'll share completely new content about the unique challenges and recipes for creating the best edge models Hope you enjoy it!
β View original post on X β @maximelabonne, 2026-04-07 11:11 UTC
-

New Scaling Laws for 350M Model Training Tokens
By
–
FACT: If you don't train your 350M model on 28T tokens, you're not optimal Nicholas Roberts (@nick11roberts) That new LFM2.5-350M is super overtrained, right? And everyone was shocked about how far they pushed it? As it turns out, we have a brand new scaling law for that! π§΅ [1/n] β https://nitter.net/nick11roberts/status/2041141606305124486#m
β View original post on X β @maximelabonne, 2026-04-06 15:05 UTC
-

LFM-Zero: Foundation Model Trained on Zero Tokens
By
–
Massive unlock: pre-training for $0 Alexander Amini (@xanamini) Three years ago we started working on a stealth project that we werenβt sure weβd ever talk about publicly… until today. Breakthrough: Introducing LFM-Zero: the first foundation model trained on 0 tokens. No pretraining. No finetuning. No data. Instead, we initialize from an implicit probabilistic prior over the underlying data-generating process, allowing the model to converge without ever observing data. LFM-Zero matches or surpasses models trained on 10T+ tokens across reasoning, coding, and multimodal tasks. Turns out that pretraining was just regularization that was holding us back. > Read our Tech Report here: tinyurl.com/lfm-zero β https://nitter.net/xanamini/status/2039403320247480469#m
β View original post on X β @maximelabonne, 2026-04-01 22:11 UTC
-
What if AI Chat Interfaces Had Gravity
By
–
What if AI chat interfaces had gravity? pic.twitter.com/CgaZbIcjzF
— Xenova (@xenovacom) 1 avril 2026What if AI chat interfaces had gravity?
β View original post on X β @maximelabonne, 2026-04-01 16:02 UTC
-
LFM2.5-350M: Tiny Agentic AI Model Running in Browser
By
–
This is an always-on model living in your browser.
— Maxime Labonne (@maximelabonne) 31 mars 2026
Sub-500 MB QA, data extraction, tool use πͺ https://t.co/6ttaJBTxAcThis is an always-on model living in your browser. Sub-500 MB QA, data extraction, tool use πͺ Xenova (@xenovacom) NEW: LiquidAI just released LFM2.5-350M, a tiny model that brings agentic AI and tool-calling capabilities to resource-constrained environments. π€― It can even run locally in your browser via WebGPU, serving as a powerful companion while you browse the web. Try the demo! π β https://nitter.net/xenovacom/status/2039043406823833964#m
β View original post on X β @maximelabonne, 2026-03-31 21:46 UTC
-

Liquid AI Releases LFM2.5-350M Compact Agentic Model
By
–
Very proud of our tiny powerhouse. Amazing performance in data extraction and tool use at such a small scale. Enjoy! π Liquid AI (@liquidai) Today, we release LFM2.5-350M. Agentic loops at 350M parameters. A 350M model trained for reliable data extraction and tool use, where models at this scale typically struggle. <500MB when quantized, built for environments where compute, memory, and latency are constrained. π§΅ β https://nitter.net/liquidai/status/2039029358224871605#m
β View original post on X β @maximelabonne, 2026-03-31 17:29 UTC
-
24B Model Running in Browser with WebGPU Technology
By
–
24B model running in your browser! https://t.co/dorx4Ux2x7
— Maxime Labonne (@maximelabonne) 25 mars 202624B model running in your browser! Xenova (@xenovacom) WebGPU is INSANE! π€― Here's a 24B parameter model running locally in a web browser, at a blazing ~50 tokens/second on my M4 Max. β‘οΈ It's the largest model we've ever run with Transformers.js… and we're not stopping here. Big announcement soon. β https://nitter.net/xenovacom/status/2036908326462665211#m
β View original post on X β @maximelabonne, 2026-03-25 21:10 UTC
-

Hack #05: AI in Space Hackathon Registration Now Open
By
–
AI is beginning to move beyond the cloudsβ¦ Registration is open for Hack #05: AI in Space, in collaboration with @DPhiSpace. A hackathon exploring what becomes possible when AI operates closer to satellites, orbital systems, and space-based data. For developers, researchers, and builders interested in the future of AI in space. Register β luma.com/n9cw58h0 Learn more β hackathons.liquid.ai π Join the conversation β discord.com/channels/1385439β¦
β View original post on X β @maximelabonne, 2026-03-19 15:04 UTC
-

LFM2 Models Dominate Fine-Tuneability Benchmark Among Small LLMs
By
–
Very interesting results about the fine-tunability of different models. π LFM2 is more flexible than alternatives. It also confirms some common knowledge about RL degrading fine-tuneability. Jacek Golebiowski (@j_golebiowski) We benchmarked 15 small language models across 9 tasks to find out which one you should actually fine-tune. The most surprising result: Liquid AI's LFM2-350M ranked #1 for tunability. 350M parameters, absorbing training signal more effectively than models 20x its size. The entire LFM2 family swept the top 3 spots. No other architecture came close. LFM2-350M: avg rank 2.11 (Β±0.89) LFM2-1.2B: avg rank 3.44 LFM2.5-1.2B-Instruct: avg rank 4.89 That tight CI means it's consistent across every task type, not just a few lucky benchmarks. β https://nitter.net/j_golebiowski/status/2033611679645266280#m
β View original post on X β @maximelabonne, 2026-03-16 19:06 UTC