We hire a developer to build parsers for a complicated file format. It takes a week per parser. Gemini 3 is the first LLM that is able to create a parser from scratch, and it does it very well. Within a minute, 1-shot-right. I am blown away.
To this day, I still don't understand why Claude gets more acclaim for coding. Gemini 2.5 consistently outperformed Claude and ChatGPT mostly because of the much larger context.
I use Gemini cli, Claude Code and Codex daily. If I present the same bug to all 3, Gemini often is the one missing a part of the solution or drawing the wrong conclusion. I am curious for G3.
I truly do not see the USP for Mistral other than being based in EU. It's former USP of setting up their models on-premises for clients is now moot with the proliferation of open frontier models. I'd love to be proven wrong but I don't see a path forward for Mistral at this point, given how far they're behind and their overall lack of competitive advantages for an AI Lab like access to hardware, cheap energy or a mass of AI talent.
They’ve built performance, enterprise utility, privacy, sovereignty, open innovation and strategic partnerships into their core story. It's quite a list. The models are opensource, Voxtral outperforms Whisper in terms of accuracy.
Those Qwen3 2507 models are the local creme-de-la-creme right now. If you've got any sort of GPU and ~32gb of RAM to play with, the A3B one is great for pair-programming tasks.
They mean: You server will get connection timeouts through CloudFlare proxy