This ML Prof told me that the hallucination rate for frontier reasoning LLMs is “next to nil” And then gave me data, only after I pushed him, showing a best-case rate of 4.6% (which of course is benchmark specific). 4.6% is not “next to nil”. Imagine if your accountant hallucinated 4.6% of the time. Or worse, your pilot. Aran Nayebi (@aran_nayebi) Have you had a chance to try the latest reasoning models? You'll see their hallucination rate is next to nil. In fact, there’s a big difference between frontier reasoning models & the base LLMs that're freely available to the public, see e.g. here: nitter.net/aran_nayebi/status/202… — https://nitter.net/aran_nayebi/status/2041249684698648922#m
→ View original post on X — @garymarcus, 2026-04-06 22:25 UTC
Leave a Reply