Skip to content
Back to Blog
Model Releases

Mistral Small 3.2: The 24B Open-Source Reasoning King

Mistral AI releases Small 3.2 with improved instruction following and Apache 2.0 licensing for enterprise developers.

June 10, 2025
Model ReleaseMistral Small 3.2
Mistral Small 3.2 - official image

Introduction

Mistral AI has officially unveiled its latest iteration in the Small series, the Mistral Small 3.2, on June 10, 2025. This release represents a significant milestone in the open-weight model landscape, specifically targeting developers who require high-performance reasoning without the overhead of massive frontier models. Following the success of the previous 3.1 version, this update focuses heavily on consolidating reasoning capabilities, instruction following, and coding efficiency into a single, versatile architecture.

The industry has been watching Mistral closely as they challenge the dominance of closed-source giants with their open-weight strategy. The Small 3.2 is not just an incremental update; it is a strategic pivot towards enterprise-grade efficiency. By maintaining a 24B parameter count while improving instruction adherence, Mistral aims to simplify the AI stack for organizations that previously jumbled multiple models for different tasks.

  • Release Date: 2025-06-10
  • Update: Mistral Small 3.1
  • License: Apache 2.0

Key Features & Architecture

The architecture of Mistral Small 3.2 is designed for density and efficiency. It utilizes a Mixture of Experts (MoE) structure that allows the model to activate only the necessary parameters for specific tasks, reducing inference costs significantly compared to dense models of similar size. The context window has been expanded to support long-form documentation and complex agent workflows, ensuring that the model can maintain coherence over extended inputs.

Developers will appreciate the Apache 2.0 license, which removes the restrictive commercial limitations found in many proprietary models. This license allows for unrestricted deployment, fine-tuning, and redistribution, making it ideal for proprietary enterprise applications. The model integrates multimodal capabilities, allowing it to process text and code simultaneously, which is crucial for modern agentic workflows.

  • Parameters: 24B
  • Context Window: 128k tokens
  • Architecture: MoE
  • Multimodal: Yes

Performance & Benchmarks

In terms of raw capability, Mistral Small 3.2 shows marked improvements over its predecessor in logical reasoning and coding tasks. On the MMLU benchmark, it achieves a score of 84.5, surpassing the 82.0 score of the Small 3.1 version. This indicates a stronger grasp of general knowledge and reasoning. For developers focused on software engineering, the HumanEval benchmark score of 88.2 demonstrates its proficiency in generating functional code snippets without the need for extensive prompting.

The model also excels on the SWE-bench repository, where it successfully resolves complex issues in 35% of cases, a significant jump from the previous generation. These benchmarks confirm that the 24B parameter count is being utilized more effectively, delivering frontier-level performance at a fraction of the inference cost of larger models. This efficiency is key for real-time applications where latency matters.

  • MMLU: 84.5
  • HumanEval: 88.2
  • SWE-bench: 35% pass rate
  • Latency: <50ms (local)

API Pricing

Mistral AI has structured the pricing for the Small 3.2 API to be highly competitive for high-volume applications. The input cost is set at $0.15 per million tokens, while the output cost is $0.45 per million tokens. This pricing model is designed to encourage experimentation and long-running agent interactions without breaking the budget. For teams processing millions of tokens daily, this translates to substantial cost savings compared to proprietary alternatives.

There is also a free tier available for developers and hobbyists, offering up to 100,000 tokens per month at no cost. This tier is perfect for prototyping and testing integrations. The value comparison is clear: for every dollar spent, developers get significantly more tokens and higher quality responses than with comparable open-source models hosted on other platforms.

  • Free Tier: 100k tokens/month
  • Input Price: $0.15 / 1M
  • Output Price: $0.45 / 1M
  • No hidden fees

Comparison Table

When evaluating the market landscape, Mistral Small 3.2 stands out for its balance of cost and capability. Compared to Llama 3.1 70B, it offers similar reasoning at a lower parameter count, making it faster to run on consumer hardware. Against Gemma 3, it provides better instruction following, and compared to Phi-4, it offers a larger context window for enterprise data. The table below details the specific technical metrics that define its competitive edge in the current 2025 AI market.

  • Best for: Reasoning & Cost
  • Context: 128k
  • License: Open

Use Cases

The versatility of Mistral Small 3.2 makes it suitable for a wide range of applications. It is particularly well-suited for coding assistants, where its ability to understand complex codebases is essential. Developers can integrate it into IDE plugins to provide real-time refactoring suggestions or bug fixes. Additionally, its reasoning capabilities make it ideal for customer support agents that need to handle nuanced queries without hallucinating information.

For RAG (Retrieval-Augmented Generation) systems, the 128k context window allows the model to ingest entire documentation sets or code repositories. This reduces the need for complex chunking strategies. Enterprise teams can deploy this model on-premise or via private cloud, ensuring data sovereignty while leveraging the latest open-weight technology.

  • Coding Assistants
  • Customer Support Agents
  • RAG Systems
  • On-Premise Deployment

Getting Started

Accessing Mistral Small 3.2 is straightforward for developers. The model is available via the official Mistral API, which supports standard REST endpoints and Python SDKs. You can sign up for an API key on the Mistral platform to start making requests immediately. For those who prefer local deployment, the weights are hosted on Hugging Face, allowing for easy integration into local environments using standard inference libraries like vLLM or Ollama.

Documentation is comprehensive, providing examples for Python, JavaScript, and Go. The GitHub repository contains fine-tuning scripts and benchmark results to help developers optimize the model for their specific hardware constraints. With the Apache 2.0 license, you have the freedom to build, modify, and distribute your applications using this model without restriction.

  • API Endpoint: https://api.mistral.ai
  • SDK: Python, JS, Go
  • Weights: Hugging Face
  • Docs: mistral.ai/docs

Comparison

API Pricing β€” Input: $0.15 / 1M tokens / Output: $0.45 / 1M tokens / Context: 128k


Sources

Mistral AI News

Mistral GitHub

Yahoo Finance Report