It's not cheaper than Deepseek V3.1, though, and Deepseek outperforms on nearly everything. And only between 1- 3x the throughput based on the openrouter metrics (near equivalent throughput if you use an FP8 quant). Wish I could be a little more excited about this one.
amai · 14h ago
Mistral seems to be the only company that doesn‘t fake benchmarks. That makes it of course not exciting. But it doesn‘t have to be to be useful.
Love the information on this guy's website, dislike the layout of information and use of Google Sheets on pages. Makes it difficult to navigate and grok.
boramalper · 20h ago
I guess this one (Mistral Medium 3) won't be open?
If you carefully read their performance chart, they beat every open source model they bothered to list.
So if we were in a world in which they released this model, and only the listed open source models existed, they would be le SOTA!
ed · 12h ago
DeepSeek wins most benchmarks according to that chart, so not quite SOTA, and there's no mention of model size so it's hard to compare efficiency
qeternity · 11h ago
> There's no reason for someone to use this over Anthropic, OpenAI, etc
Uh except that you can host this yourself? You just have to license the model. What is the issue?
moralestapia · 9h ago
>You just have to license the model.
That's the issue.
deanc · 14h ago
They are dropping some hints about a larger model and something relating to open models in the final paragraph:
> With the launches of Mistral Small in March and Mistral Medium today, it’s no secret that we’re working on something ‘large’ over the next few weeks. With even our medium-sized model being resoundingly better than flagship open source models such as Llama 4 Maverick, we’re excited to ‘open’ up what’s to come :)
kristianp · 13h ago
The only hint at its size is that it requires "self-hosted environments of four GPUs and above".
bn-l · 16h ago
The medium model is crazy fast. Reminds me of maverick on groq (except better according to their own testing).
There's no reason for someone to use this over Anthropic, OpenAI, etc ... they don't even outperform them.
You cant do that with the providers you listed.
They event hint at it on their PR:
"Mistral La Plateforme and Amazon Sagemaker, and soon on IBM WatsonX, NVIDIA NIM, Azure AI Foundry, and Google Cloud Vertex"
My point being, why would you pay to use a Mistral model hosted on Azure, instead of using any other company model hosted on Azure?
My answer to that, yesterday, would have been "because the model is free and unrestricted, I only pay for hardware"; that premise is gone today.
They released it at the same time of the article we are discussing on...
https://mistral.ai/news/le-chat-enterprise
If you carefully read their performance chart, they beat every open source model they bothered to list.
So if we were in a world in which they released this model, and only the listed open source models existed, they would be le SOTA!
Uh except that you can host this yourself? You just have to license the model. What is the issue?
That's the issue.
> With the launches of Mistral Small in March and Mistral Medium today, it’s no secret that we’re working on something ‘large’ over the next few weeks. With even our medium-sized model being resoundingly better than flagship open source models such as Llama 4 Maverick, we’re excited to ‘open’ up what’s to come :)