BREAKING: New bartowski Gemma-4 26B-A4B-it MoE GGUF Just Dropped 🤯 Dropped the IQ4_NL GGUF of Google’s Gemma-4 26B-A4B-it MoE (~26B total / ~4B active)! Bartowski’s GGUF Setup: 👇🏻 🧠Revised & quantized with llama.cpp imatrix 🚀Quant: gemma-4-26B-A4B-it-IQ4_NL.gguf 14.70 GB 💻Full 256K context window Native (text + vision) MoE Performance Wins: 👇🏻 🏆 Efficient Mixture-of-Experts (8 active / 128 total) 🚨 Clean, accurate tool calls with no overthinking 🤯 Noticeably stronger agentic workflow MoEs (beats Qwen 3.5-35B-A3B in tool-use precision) 🤖 Built for local Hermes Agent / agentic meta Quants from IQ4-XS to Q4_K_M:👇🏻 🔥IQ4_NL (14.70 GB) — best accuracy/speed balance 🔥IQ4_XS (~14.2 GB) — lightest high-quality option 🔥Q4_K_S (~15.8 GB) Q4_K_M (~17 GB) Try It 👇🏻 huggingface.co/bartowski/goo…
→ View original post on X — @huggingface, 2026-04-03 20:36 UTC

Leave a Reply