Mistral AI’s latest announcement introduces DevStral 2 (123B parameters), DevStral Small 2 (24B), and the Mistral Vibe CLI, a terminal-native coding assistant built for agentic coding tasks. Both models are fully open source and tuned for production workflows, while the new Vibe CLI brings project-aware editing, code search, version control, and execution directly into the terminal.
Together, these updates aim to speed up developer workflows by making large-scale code refactoring, bug fixes, and feature development more automated, and in this guide we’ll outline the technical capabilities of each tool and provide hands-on examples to get started.
DevStral 2 is a 123-billion-parameter dense transformer designed specifically for software engineering agents. It features a 256K-token context window, enabling it to analyze entire code repositories at once. Despite its size, it is much smaller than competitor models: for example, DevStral 2 is 5x smaller than DeepSeek v3.2 and 8x smaller than Kimi K2 yet matches or exceeds their performance. This compactness makes DevStral 2 practical for enterprise deployment.
The Key technical highlights of DevStral 2 include:

These capabilities mean DevStral 2 is not just a powerful code completion model, but a true coding assistant that maintains state across an entire project. For developers, this translates to faster, more reliable automated changes: for example, DevStral 2 can understand a project’s file structure and dependencies, propose code modifications across many modules, and even apply fixes iteratively if tests fail.
You can learn more about the pricing of DevStral 2 from their official page.


Utilize Mistral’s official SDK to submit coding requests. For example, if you want DevStral 2 to redo a Python function for better speed, you can type:
!pip install mistralai
from mistralai import Mistral
import os
from getpass import getpass
api_key = getpass("Enter your Mistral API Key: ")
client = Mistral(api_key=api_key)
response = client.chat.complete(
model="devstral-2512", # correct model name
messages=[
{"role": "system", "content": "You are a Python code assistant."},
{"role": "user", "content": (
"Refactor the following function to improve performance:\n"
"```python\ndef compute_numbers(n):\n"
" result = []\n"
" for i in range(n):\n"
" if i % 100 == 0:\n"
" result.append(i**2)\n"
" return result\n```"
)}
]
)
print(response.choices[0].message.content)
The request is made to DevStral 2 to make a loop function faster. The AI will examine the function and give a reformed version (for instance, recommending using list comprehensions or vectorized libraries). Although the Python SDK makes it easier to interact with the model, you may also opt to make HTTP requests for direct API access if that is your choice.

Hugging Face has DevStral 2 weights available meaning that it is possible to run the model locally (if your hardware is good enough) using the Transformers library. Just to give an example:
!pip install transformers # make sure you have transformers installed
# optionally: pip install git+https://github.com/huggingface/transformers if using bleeding-edge
from transformers import MistralForCausalLM, MistralCommonBackend
import torch
model_id = "mistralai/Devstral-2-123B-Instruct-2512"
# Load tokenizer and model
tokenizer = MistralCommonBackend.from_pretrained(model_id, trust_remote_code=True)
model = MistralForCausalLM.from_pretrained(model_id, device_map="auto", trust_remote_code=True)
# Optionally, set dtype for better memory usage (e.g. bfloat16 or float16) if you have GPU
model = model.to(torch.bfloat16)
prompt = (
"Write a function to merge two sorted lists of integers into one sorted list:\n"
"```python\n"
"# Input: list1 and list2, both sorted\n"
"```"
)
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
The displayed code snippet utilizes the “DevStral 2 Instruct” model to produce a complete Python function similar to the previous code.
DevStral Small 2 brings the same design principles to a much smaller model. It has 24 billion parameters and the same 256K context window but is sized to run on a single GPU or even a high-end consumer CPU.
The Key attributes of DevStral Small 2 include:

From a developer’s perspective, DevStral Small 2 enables fast prototyping and on-device privacy. Because inference is quick (even running on CPU), you get tight feedback loops when testing changes. And since the runtime is local, sensitive code never has to leave your infrastructure.
Just like DevStral 2, the Small model is available via the Mistral API. In the Python SDK, you could do:
!pip install mistralai
from mistralai import Mistral
import os
from getpass import getpass
api_key = getpass("Enter your Mistral API Key: ")
client = Mistral(api_key=api_key)
response = client.chat.complete(
model="devstral-small-2507", # updated valid model name
messages=[
{"role": "system", "content": "You are a Python code assistant."},
{"role": "user", "content": (
"Write a clean and efficient Python function to find the first "
"non-repeating character in a string. Return None if no such "
"character exists."
)}
]
)
print(response.choices[0].message.content)
Output:

Mistral Vibe CLI is an open-source, Python-based command-line interface that turns DevStral into an agent running in your terminal. It provides a conversational chat interface that understands your entire project. Vibe automatically scans your project’s directory and Git status to build context.
You can reference files with @autocompletion, execute shell commands with exclamation(!) , and use slash commands ( /config, /theme, etc.) to adjust settings. Because Vibe can “understand your entire codebase and not just the file you’re editing”, it enables architecture-level reasoning (for example, suggesting consistent changes across modules).
The main characteristics of Vibe CLI are the following:
grep), version control, and running shell commands. For instance, it can read a file with the read_file command, apply a patch by writing it to the file with the write_file command, search for the repo using grep, etc. /config) can be used through /slash. This results in a seamless CLI experience, complete with persistent history and even customization of the theme. --prompt or piping) to script batch tasks for scripting. You can create a config.toml file to set the default models (e.g. pointing to DevStral 2 via API), switch --auto-approve on or off for tool execution, and limit risky operations in sensitive repos. uv tool install mistral-vibe
OR
curl -LsSf https://mistral.ai/vibe/install.sh | sh
OR
pip install mistral-vibe
Vibe

Prompt: vibe "Write a Python function to reverse a linked list"

Prompt for programmatic mode:
vibe -p "Generate a SQL schema for an employee database"


The response was satisfactory.
DevStral 2, its smaller variant, and the Mistral Vibe CLI push hard toward autonomous coding agents, giving developers faster iteration, better code insight, and lower compute costs. DevStral 2 handles multi-file code work at scale, DevStral Small 2 brings similar behavior to local setups, and Vibe CLI makes both models usable directly from your terminal with smart, context-aware tools.
To try them out, grab a Mistral API key, test the models through the API or Hugging Face, and follow the recommended settings in the docs. Whether you’re building codebots, tightening CI, or speeding up daily coding, these tools offer a practical entry into AI-driven development. Whereas DevStral 2 model series is competing in the LLM competition that’s out there, Mistral Vibe CLI is there to offer an alternative to the other CLI alternatives out there.
A. They speed up coding by enabling autonomous code navigation, refactoring, debugging, and project-aware assistance directly in the terminal.
A. DevStral 2 is a larger, more powerful model, while Small 2 offers similar agentic behavior but is light enough for local use.
A. Get a Mistral API key, explore the models through the API or Hugging Face, and follow the recommended settings in the official documentation.