Skip to content
Back to Blog
Model Releases

Mistral Small 4: The Unified Frontier Model Released

Mistral AI releases Small 4, a 119B MoE model unifying reasoning, coding, and vision into a single Apache 2.0 licensed architecture with 256K context.

March 16, 2026
Model ReleaseMistral Small 4
Mistral Small 4 - official image

Introduction: A New Era of Unified AI

Mistral AI has officially unveiled Mistral Small 4 on March 16, 2026, marking a significant shift in the open-source landscape. This release consolidates previously separate specialized models into a single, highly efficient architecture. Previously, developers had to juggle Magistral for reasoning, Pixtral for vision, and Devstral for coding tasks. Small 4 eliminates this fragmentation by offering a unified solution that handles instruct, reasoning, coding, and multimodal inputs simultaneously.

The importance of this release cannot be overstated for enterprise developers and AI engineers. By reducing the model stack complexity, organizations can lower inference costs and simplify their deployment pipelines. The model is designed to be hardware-efficient, making it suitable for cost-sensitive use cases while maintaining frontier-level performance. This strategic move positions Mistral as a dominant force in sovereign AI for European and global enterprises.

Unlike previous iterations that required separate fine-tuning for different capabilities, Small 4 features a configurable reasoning parameter. This allows users to dynamically adjust the computational depth required for complex tasks without switching models. The Apache 2.0 license ensures maximum flexibility for commercial and research applications, fostering a more open ecosystem for AI innovation.

  • Release Date: 2026-03-16
  • License: Apache 2.0
  • Unified Capabilities: Reasoning, Vision, Coding, Chat

Key Features & Architecture

Under the hood, Mistral Small 4 utilizes a sophisticated Mixture of Experts (MoE) architecture. The model boasts 119 billion total parameters, with 6.5 billion active parameters during inference. This design choice ensures that the model remains lightweight enough for efficient deployment while retaining the capacity to handle complex, multi-step reasoning tasks that typically require larger dense models.

The context window has been expanded significantly to 256K tokens, allowing the model to process entire codebases or lengthy documents in a single pass. This is a crucial upgrade for Retrieval Augmented Generation (RAG) applications where long-context understanding is paramount. The multimodal capabilities are integrated natively, meaning vision tasks do not require separate image encoders, reducing latency and improving throughput.

The architecture is optimized for hardware efficiency, competing directly with recent releases from OpenAI and Anthropic. It supports configurable reasoning levels, enabling users to balance cost and performance dynamically. This flexibility is key for agentic workflows where the model must decide when to invoke deeper reasoning chains versus simple instruction following.

  • Parameters: 119B Total (6.5B Active)
  • Context Window: 256K Tokens
  • Architecture: MoE (Mixture of Experts)

Performance & Benchmarks

In terms of raw performance, Mistral Small 4 demonstrates competitive results across standard benchmarks. On MMLU, the model achieves a score of 88.5%, surpassing previous open-source variants. For coding tasks, it scores 92% on HumanEval, making it a viable replacement for specialized coding agents. The model also shows strong performance on SWE-bench, indicating its capability to solve real-world software engineering issues.

Compared to the specialized predecessors, Small 4 shows no significant degradation in performance despite the unified architecture. In multimodal benchmarks, it outperforms Pixtral in reasoning tasks involving charts and diagrams. The configurable reasoning parameter allows users to trade off latency for accuracy, achieving a Pareto optimal balance for most enterprise applications.

  • MMLU Score: 88.5%
  • HumanEval Score: 92%
  • SWE-bench Pass Rate: 65%

API Pricing & Value

Mistral AI has structured the pricing for Small 4 to be highly competitive, targeting cost-sensitive use cases. The input cost is set at $0.12 per million tokens, while the output cost is $0.36 per million tokens. This pricing structure is significantly lower than comparable proprietary models, making it ideal for high-volume inference workloads. The free tier availability is limited to development environments, encouraging widespread experimentation.

The value proposition extends beyond raw cost per token. By unifying three distinct models (Magistral, Pixtral, Devstral) into one, enterprises save on licensing, maintenance, and infrastructure overhead. The hardware-efficient design allows for deployment on standard GPUs, reducing the need for specialized clusters. This makes Small 4 an attractive option for startups and large enterprises alike.

  • Input Price: $0.12 / M tokens
  • Output Price: $0.36 / M tokens
  • Free Tier: Development Only

Comparison Table

When evaluating Mistral Small 4 against current market leaders, the trade-offs become clear. While proprietary models may offer slightly higher raw scores, Small 4 provides a better balance of cost, context, and open-source flexibility. Developers should consider their specific latency and budget requirements when choosing between these options.

  • See comparison data below for detailed specs.

Use Cases

Mistral Small 4 is best suited for a variety of advanced applications. It excels in agentic workflows where the model needs to reason about code changes and execute tasks autonomously. For RAG systems, the 256K context window allows for deep document analysis without truncation. Additionally, its multimodal capabilities make it perfect for document processing pipelines that require both text extraction and visual understanding.

In the realm of software development, Small 4 can serve as a primary coding assistant, handling complex refactoring and debugging tasks. Its Apache 2.0 license ensures that companies can integrate it into proprietary products without legal restrictions. This openness is a major driver for adoption in the European and global sovereign AI sectors.

  • Agentic Workflows
  • Long-Context RAG
  • Multimodal Document Processing

Getting Started

Accessing Mistral Small 4 is straightforward for developers. The model is available via the official Mistral API endpoint and through the Hugging Face Hub. SDKs are provided for Python, JavaScript, and Go, facilitating rapid integration into existing applications. The documentation includes comprehensive examples for streaming responses and managing the configurable reasoning parameters.

To deploy locally, developers can download the model weights directly from the repository. The hardware-efficient design means it can run on consumer-grade GPUs for inference. Mistral AI also offers a playground for testing the model's capabilities before committing to production deployment.

  • API Endpoint: api.mistral.ai
  • Platform: Hugging Face Hub
  • Languages: Python, JS, Go SDKs

Comparison

API Pricing β€” Input: $0.12 / Output: $0.36 / Context: 256K


Sources

Mistral AI Blog: Small 4 Release