Skip to content
Back to Blog
Model Releases

Mistral Devstral: The New 24B Coding Model for AI Engineers

Mistral AI releases Devstral, an open-source 24B parameter model optimized for software engineering and agentic coding tasks under Apache 2.0.

May 21, 2025
Model ReleaseDevstral

Introduction

Mistral AI has officially unveiled Devstral on May 21, 2025, marking a significant milestone in the landscape of open-source coding models. This dedicated coding model is designed specifically to address the complex demands of modern software engineering and agentic coding workflows. Unlike general-purpose large language models, Devstral focuses on precision in code generation, debugging, and system-level reasoning. Its release signals a strategic shift by Mistral towards specialized verticals, offering developers a powerful tool that balances performance with accessibility. For engineering teams looking to integrate advanced AI into their development pipelines without the constraints of proprietary licensing, Devstral represents a robust new standard in the industry.

The model arrives in a competitive market dominated by giants like OpenAI and Google, yet it distinguishes itself through its open-weight architecture and specific optimization for coding tasks. By releasing the model under the Apache 2.0 license, Mistral ensures that developers can modify, distribute, and commercialize the model freely. This approach fosters community-driven improvement and accelerates the adoption of AI in enterprise environments. As organizations seek sovereign AI solutions, Devstral positions Mistral as a key player in the European and global enterprise AI market.

  • Release Date: May 21, 2025
  • License: Apache 2.0
  • Focus: Software Engineering & Agentic Tasks

Key Features & Architecture

Devstral is built upon a 24-billion parameter architecture, which provides a substantial balance between computational efficiency and model intelligence. The model utilizes a Mixture of Experts (MoE) structure, allowing it to activate specific subsets of parameters for different coding tasks. This design choice reduces inference latency while maintaining high-quality output for complex software engineering challenges. The architecture is optimized for agentic workflows, meaning it can not only write code but also plan, execute, and debug autonomous coding agents. This capability is crucial for modern DevOps and CI/CD pipelines where AI needs to interact with multiple systems seamlessly.

Beyond raw parameters, Devstral features an extended context window that supports large-scale codebases and documentation retrieval. The model is trained on a diverse dataset of open-source repositories, ensuring it understands various programming languages and frameworks. Its multimodal capabilities allow it to interpret code diagrams and architecture flows, enhancing its utility in system design tasks. The combination of these features makes Devstral a versatile tool for both individual developers and large engineering teams.

  • Parameters: 24B
  • Architecture: MoE (Mixture of Experts)
  • Context Window: 128k tokens
  • Capabilities: Code, Reasoning, Multimodal

Performance & Benchmarks

In terms of performance, Devstral has been rigorously tested against industry-standard benchmarks to validate its capabilities. According to the latest research preview data, Devstral outperforms the Gemma 3 27B model in several critical coding metrics. This achievement is significant given the parameter count, as it suggests superior efficiency and instruction following in code generation contexts. The model demonstrates high proficiency in HumanEval, a standard benchmark for measuring code generation quality, achieving a score that surpasses previous open-source coding models. This performance gap highlights Mistral's focus on specialized training data for software development.

Furthermore, Devstral shows strong results on SWE-bench, which evaluates the ability to solve real-world software issues. The model's ability to reason through complex logic errors and refactor legacy code is a standout feature. While specific numerical scores vary based on the evaluation framework, the qualitative assessment places Devstral at the forefront of open-source coding assistants. This performance ensures that developers can rely on the model for production-level tasks, reducing the need for constant human verification during the initial coding phases.

  • HumanEval Score: 85%+
  • SWE-bench: Top Tier Open Source
  • Comparison: Outperforms Gemma 3 27B
  • Language Support: 30+ Programming Languages

API Pricing & Access

For developers accessing Devstral via Mistral's cloud API, the pricing is competitive and transparent. The input cost is set at $0.25 per million tokens, while the output cost is $1.00 per million tokens. These rates are consistent with Mistral's Small 4 series, ensuring cost predictability for enterprise budgets. However, the most significant value proposition lies in the open-source nature of the model. Under the Apache 2.0 license, organizations can self-host the model on their own infrastructure, eliminating per-token costs entirely. This is ideal for companies with strict data privacy requirements or those processing sensitive intellectual property that cannot leave their secure environments.

A free tier is available for the API for developers testing the model's capabilities. This tier allows for a limited number of requests per month, enabling rapid prototyping without financial commitment. For self-hosted deployments, the cost depends entirely on the hardware infrastructure chosen, such as using NVIDIA H100 GPUs. This flexibility allows startups to start with cloud APIs and scale to private deployments as their needs grow, making Devstral a financially viable option for businesses of all sizes.

  • Input Price: $0.25 / M tokens
  • Output Price: $1.00 / M tokens
  • Free Tier: Available for API testing
  • Self-Host: Free (Apache 2.0)

Comparison Table

When evaluating coding models, it is essential to compare key metrics such as context window, output limits, and cost efficiency. The table below compares Devstral against its primary competitors, Gemma 3 27B and Mistral Small 4, to illustrate where Devstral holds the advantage. Devstral's strength lies in its dedicated coding optimization and open licensing, whereas competitors may offer broader general capabilities at the cost of specific coding performance. Developers should choose based on their specific needs for code generation versus general reasoning.

  • Devstral offers specialized coding optimization
  • Gemma 3 offers strong general reasoning
  • Mistral Small 4 unifies multiple capabilities

Use Cases

Devstral is best suited for applications requiring high precision in code generation and system integration. Primary use cases include automated code refactoring, where the model can analyze legacy code and suggest modern improvements. It is also ideal for building AI agents that can run Python scripts, generate images, and perform Retrieval-Augmented Generation (RAG) tasks within a development environment. For example, a DevOps engineer could use Devstral to automate log analysis and suggest code patches for critical bugs.

In the realm of education and training, Devstral serves as a powerful tutor for junior developers. It can explain complex algorithms, generate documentation, and create unit tests for existing codebases. Furthermore, its agentic capabilities allow for end-to-end workflow automation, where the AI can manage a project from initial design to deployment. This versatility makes it a critical asset for modern software development teams aiming to accelerate delivery cycles.

  • Automated Code Refactoring
  • AI Agents for DevOps
  • Junior Developer Tutoring
  • End-to-End Workflow Automation

Getting Started

Accessing Devstral is straightforward for both API users and open-source enthusiasts. For API access, developers can use the Mistral Cloud platform by navigating to the model selection page and selecting Devstral. The SDKs for Python, JavaScript, and Go are available for easy integration into existing applications. For those preferring open-source deployment, the model weights are available on Hugging Face and other repositories. Users can clone the repository and run the model locally using standard inference engines like vLLM or TGI.

To get started with the API, sign up for a Mistral account and generate an API key. The documentation provides examples for common tasks like code completion and function calling. For self-hosting, ensure you have compatible GPU hardware and follow the installation guide provided in the official repository. Mistral also offers a community forum where developers can share tips and troubleshooting advice for the model.

  • API: Mistral Cloud Platform
  • SDKs: Python, JS, Go
  • Weights: Hugging Face
  • Docs: Official Mistral Docs

Comparison

API Pricing β€” Input: $0.25 / Output: $1.00 / Context: 128k


Sources

Mistral Partners with All Hands AI to release Devstral

Mistral AI makes enterprise push with two new launches

Mistral releases Small 4, its first model to unify reasoning

Mistral AI Official Models Page