Mistral's revenue numbers are impressive and growing insanely fast but the bit that's most exciting as an investor is the durability of this revenue. Mistral selects who they work with carefully to ensure they actually achieve AI transformation.
— Mistral AI for Developers (@MistralDevs) 6 février 2026
Voxtral can now directly stream audio input into text output. Perfect for:
– Live subtitles
– Language learning apps
– Note-taking tools
– And more! Made a demo for you to try directly on Hugging Face! [Translated from EN to English]
Voxtral (from @MistralAI) transcription quality is quite incredible, the way it handles the punctation and all the rest, makes transcribed audio messages so much more understandable. I implemented a few fixes in the FFT and now there is no longer a skipped tokens issue in voxtral.c
The demo on https://t.co/ozhOG8iPeH is worth a try – ignore the "No microphone found" message, clicking "Record" and allowing your browser to use a microphone fixes that. It transcribes very accurately in almost real-time. It's really impressive. https://t.co/LOJUsT2sPW
The demo on huggingface.co/spaces/mistra… is worth a try – ignore the "No microphone found" message, clicking "Record" and allowing your browser to use a microphone fixes that. It transcribes very accurately in almost real-time. It's really impressive. Mistral AI (@MistralAI) Introducing Voxtral Transcribe 2, next-gen speech-to-text models by @MistralAI. State-of-the-art transcription, speaker diarization, sub-200ms real-time latency. Details in 🧵 — https://nitter.net/MistralAI/status/2019068826097213953#m
Congrats to @MistralAI on releasing Voxtral Mini 4B Realtime! 🎉 Day-0 support in vLLM! A 4B streaming ASR model achieving <500ms latency while matching offline model accuracy, supporting 13 languages. vLLM's new Realtime API `/v1/realtime` provides audio streaming – optimized for voice assistants, live subtitles, and meeting transcription! Thanks to the close collaboration between the vLLM community and @MistralAI for making this production-grade support possible 🤝 📑Model & Usage: huggingface.co/mistralai/Vox… Mistral AI (@MistralAI) Voxtral Realtime is built for voice agents and live applications. Its natively streaming architecture delivers latency configurable to sub-200ms. And at 480ms, it stays within 1-2% WER of our offline model. We release the model as open weights under Apache 2.0. — https://nitter.net/MistralAI/status/2019068828257333466#m