logoalt Hacker News

Mistral releases Devstral2 and Mistral Vibe CLI

566 pointsby pemberyesterday at 2:45 PM277 commentsview on HN

Comments

simonwyesterday at 4:45 PM

  llm install llm-mistral
  llm mistral refresh
  llm -m mistral/devstral-2512 "Generate an SVG of a pelican riding a bicycle"
https://tools.simonwillison.net/svg-render#%3Csvg%20xmlns%3D...

Pretty good for a 123B model!

(That said I'm not 100% certain I guessed the correct model ID, I asked Mistral here: https://x.com/simonw/status/1998435424847675429)

show 7 replies
esafakyesterday at 4:27 PM

Less than a year behind the SOTA, faster, and cheaper. I think Mistral is mounting a good recovery. I would not use it yet since it is not the best along any dimension that matters to me (I'm not EU-bound) but it is catching up. I think its closed source competitors are Haiku 4.5 and Gemini 3 Pro Fast (TBA) and whatever ridiculously-named light model OpenAI offers today (GPT 5.1 Codex Max Extra High Fast?)

show 2 replies
InsideOutSantayesterday at 7:38 PM

I gave Devstral 2 in their CLI a shot and let it run over one of my smaller private projects, about 500 KB of code. I asked it to review the codebase, understand the application's functionality, identify issues, and fix them.

It spent about half an hour, correctly identified what the program did, found two small bugs, fixed them, made some minor improvements, and added two new, small but nice features.

It introduced one new bug, but then fixed it on the first try when I pointed it out.

The changes it made to the code were minimal and localized; unlike some more "creative" models, it didn't randomly rewrite stuff it didn't have to.

It's too early to form a conclusion, but so far, it's looking quite competent.

show 1 reply
embedding-shapeyesterday at 4:09 PM

Look interesting, eager to play around with it! Devstral was a neat model when it released and one of the better ones to run locally for agentic coding. Nowadays I mostly use GPT-OSS-120b for this, so gonna be interesting to see if Devstral 2 can replace it.

I'm a bit saddened by the name of the CLI tool, which to me implies the intended usage. "Vibe-coding" is a fun exercise to realize where models go wrong, but for professional work where you need tight control over the quality, you can obviously not vibe your way to excellency, hard reviews are required, so not "vibe coding" which is all about unreviewed code and just going with whatever the LLM outputs.

But regardless of that, it seems like everyone and their mother is aiming to fuel the vibe coding frenzy. But where are the professional tools, meant to be used for people who don't want to do vibe-coding, but be heavily assisted by LLMs? Something that is meant to augment the human intellect, not replace it? All the agents seem to focus on off-handing work to vibe-coding agents, while what I want is something even tighter integrated with my tools so I can continue delivering high quality code I know and control. Where are those tools? None of the existing coding agents apparently aim for this...

show 8 replies
freakynittoday at 2:57 AM

So I tested the bigger model with my typical standard test queries which are not so tough, not so easy. They are also some that you wouldn't find extensive training data for. Finally, I already have used them to get answers from gpt-5.1, sonnet 4.5 and gemini 3 ....

Here is what I think about the bigger model: It sits between sonnet 4 and sonnet 4.5. Something like "sonnet 4.3". The response sped was pretty good.

Overall, I can see myself shifting to this for reguar day-to-day coding if they can offer this for copetitive pricing.

I'll still use sonnet 4.5 or gemini 3 for complex queries, but, for everything else code related, this seems to be pretty good.

Congrats Mistral. You most probably have caught up to the big guys. Not there yet exactly, but, not far now.

pluralmonadyesterday at 4:07 PM

I'm sure I'm not the only one that thinks "Vibe CLI" sounds like an unserious tool. I use Claude Code a lot and little of it is what I would consider Vibe Coding.

show 5 replies
princehonestyesterday at 6:12 PM

Let's say you had a hardware budget of $5,000. What machine would you buy or build to run Devstral Small 2? The HuggingFace page claims it can run on a Mac with 32 GB of memory or an RTX 4090. What kind of tokens per second would you get on each? What about DGX Spark? What about RTX 5090 or Pro series? What about external GPUs on Oculink with a mini PC?

show 3 replies
eavan0yesterday at 8:11 PM

I'm glad it's not another LLM CLI that uses React. Vibe-cli seems to be built with https://github.com/textualize/textual/

show 1 reply
zimbatmyesterday at 5:32 PM

Just added it to our inventory. For those of you using Nix:

    nix run github:numtide/llm-agents.nix#mistral-vibe
The repo is updated daily.
show 1 reply
pzmarzlyyesterday at 4:26 PM

10x cheaper price per token than Claude, am I reading it right?

As long as it doesn't mean 10x worse performance, that's a good selling point.

show 3 replies
rubin55yesterday at 8:37 PM

This is great! I just made an AUR package for it: https://aur.archlinux.org/packages/mistral-vibe

alexmorleyyesterday at 4:20 PM

Does anyone know where their SWE-bench Verified results are from? I can't find matching results on the leaderboards for their models or the Claude models and they don't provide any links.

rsolvayesterday at 7:47 PM

Ah, finally! I was checking just a few days ago if they had a Claude Code-like tool as I would much rather give money to a European effort. I'll stop my Pro subscription at Anthropic and switch over and test it out.

SyneRyderyesterday at 6:04 PM

I was briefly excited when Mistral Vibe launched and mentions "0 MCP Servers" in its startup screen... but I can't find how to configure any MCP servers. It doesn't respond to the /mcp command, and asking Devstral 2 for help, it thinks MCP is "Model Context Preservation". I'd really like to be able to run my local MCP tools that I wrote in Golang.

I'm team Anthropic with Claude Max & Claude Code, but I'm still excited to see Mistral trying this. Mistral has occasionally saved the day for me when Claude refused an innocuous request, and it's good to have alternatives... even if Mistral / Devstral seems to be far behind the quality of Claude.

show 1 reply
joostdevriesyesterday at 5:18 PM

Very nice that there's a coding cli finally. I have a Mistral Pro account. I hope that it will be included. It's the main reason to have a Pro account tbh.

mentalgearyesterday at 9:43 PM

Just tried it out via their free API and the Roo Code VSCode extension, and it's impressive. It walked through a data analytics and transformation problem (150.000 dataset entries) I have been debugging for the past 2 hours.

pshirshovyesterday at 9:23 PM

> Mistral Code is available with enterprise deployments. > Contact our team to get started.

The competition is much smoother. Where are the subscriptions which would give users the coding agent and the chat for a flat fee and working out of the box?..

weitendorfyesterday at 6:56 PM

Open sourcing the TUI is pretty big news actually. Unless I missed something, I had to dig a bit to find it, but I think this is it: https://github.com/mistralai/mistral-vibe

Going to start hacking on this ASAP

syntaxingyesterday at 8:05 PM

Extremely happy with this release, the previous Devstral was great but training it for open hands crippled the usefulness. Having their own CLI dev tool will hopefully be better

show 1 reply
tucnakyesterday at 4:03 PM

I'm so glad Mistral never sold out. We're really lucky to have them in the EU at the time when we're so focused on mil-tech etc.

show 2 replies
therealmarvyesterday at 5:07 PM

offtopic but it hurts my eyes: I dislike for their font choice and their "cool looks" in their graphics.

Surprising and good is only: Everything including graphics fixed when clicking my "speedreader" button in Brave. So they are doing that "cool look" by CSS.

show 1 reply
rwkyyesterday at 6:54 PM

I gave it the job of modifying a fairly simple regex replacement and it took a while over 5 minutes, claude failed on the same prompt (which surprised me), codex did a similar job but faster. So all in all not bad!

maelitoyesterday at 8:06 PM

Finally, we can use a european model to replace claude code.

badsectoraculayesterday at 4:26 PM

> Devstral 2 ships under a modified MIT license, while Devstral Small 2 uses Apache 2.0. Both are open-source and permissively licensed to accelerate distributed intelligence.

Uh, the "Modified MIT license" here[0] for Devstral 2 doesn't look particularly permissively licensed (or open-source):

> 2. You are not authorized to exercise any rights under this license if the global consolidated monthly revenue of your company (or that of your employer) exceeds $20 million (or its equivalent in another currency) for the preceding month. This restriction in (b) applies to the Model and any derivatives, modifications, or combined works based on it, whether provided by Mistral AI or by a third party. You may contact Mistral AI ([email protected]) to request a commercial license, which Mistral AI may grant you at its sole discretion, or choose to use the Model on Mistral AI's hosted services available at https://mistral.ai/.

[0] https://huggingface.co/mistralai/Devstral-2-123B-Instruct-25...

show 4 replies
tigranbsyesterday at 7:56 PM

Somehow it writes bad React code and misses to check linting prompts half the time. But surprisingly, the Python coding was great!

whimsicalismyesterday at 4:34 PM

> Model Size (B tokens)

How is that a measure of model size? It should either be parameter size, activated parameters, or cost per output token.

Looks like a typo because the models line up with reported param sizes.

Poudlardoyesterday at 5:41 PM

will definetey try mistral vibe with gpt-oss-20b

qwertoxyesterday at 5:52 PM

Let's see which company becomes the first to sell "coding appliances": hardware with a model good enough for normal coding.

If Mistral is so permissive they could be the first ones, provided that hardware is then fast/cheap/efficient enough to create a small box that can be placed in an office.

Maybe in 5 years.

show 4 replies
tgtweakyesterday at 10:46 PM

PSA: 10X savings when you have to prompt it 10 times to get the correct solution is not actually faster.

kevin061yesterday at 4:25 PM

I am very disappointed they don't have an equivalent subscription for coding to the 200 EUR ChatGPT or Claude one, and it is only available for Enterprise deployments.

The only thing I found is a pay-as-you-go API, but I wonder if it is any good (and cost-effective) vs Claude et al.

show 2 replies
cyp0633yesterday at 4:16 PM

In a figure: Model size (B tokens)?

abusonyesterday at 6:37 PM

did anyone test how up to date is knowledge?

After querying the model about .NET, it seems that its knowledge comes from around June 2024.

show 1 reply
moffkalastyesterday at 6:37 PM

Looks like another Deepseek distil like the new Ministrals. For every other use case that would be an insult, but for coding that's a great approach given how much lead in coding performance Qwen and Deepseek have on Mistral's internal datasets. The Small 24B seems to have a decent edge on 30BA3B, though it'll be comparatively extremely slow to run.

da_grift_shiftyesterday at 4:59 PM

Can Vibe CLI help me vibe code PRs for when I vibe on the https://github.com/buttplugio/buttplug repo?

show 1 reply
jedisct1yesterday at 4:41 PM

Yet another CLI.

Why does every AI provider need to have its own tool, instead of contributing to existing tools like Roo Code or Opencode?

show 1 reply
villgaxyesterday at 6:30 PM

Modified MIT?????

Just call it Mistral License & flush it down