Skip to content
Back to Blog
Model Releases

Mistral Medium 3: The Open-Source GPT-4o Challenger

Mistral AI unveils Mistral Medium 3, a front-tier Apache 2.0 model competing with GPT-4o, featuring strong multilingual support and edge deployment capabilities.

May 14, 2025
Model ReleaseMistral Medium 3
Mistral Medium 3 - official image

Introduction

Mistral AI has officially released Mistral Medium 3 on May 14, 2025, marking a significant milestone in open-weight artificial intelligence. This new model is designed to bridge the gap between enterprise-grade performance and accessible, open-source innovation. Unlike previous iterations locked behind proprietary walls, Mistral Medium 3 positions itself as a direct competitor to GPT-4o while maintaining full transparency.

The release comes amidst a surge in demand for lightweight, high-performance models that can operate on edge devices without sacrificing reasoning capabilities. By leveraging a Mixture of Experts (MoE) architecture, Mistral aims to provide a solution that scales efficiently across various hardware configurations, from high-end data centers to consumer laptops.

  • Released: May 14, 2025
  • License: Apache 2.0
  • Architecture: MoE (Mixture of Experts)

Key Features & Architecture

Mistral Medium 3 introduces a robust architecture designed for efficiency and versatility. The model utilizes a sparse MoE structure, allowing it to activate only a subset of parameters for specific tasks, thereby reducing inference latency. This design choice is critical for developers looking to deploy models on resource-constrained environments like drones or edge servers.

Multilingual capabilities are a standout feature, with native support for over 100 languages. This ensures that applications built on Mistral Medium 3 can serve global audiences without significant localization overhead. The Apache 2.0 license further solidifies its position in the open-source ecosystem, permitting commercial use without restrictive clauses.

  • 128K Context Window
  • 100+ Native Languages
  • Apache 2.0 License
  • Sparse MoE Architecture

Performance & Benchmarks

In independent evaluations, Mistral Medium 3 demonstrates competitive performance against established closed-source models. On the MMLU benchmark, the model achieves a score of 85.4, closely trailing GPT-4o but surpassing previous open-weight baselines. For coding tasks, HumanEval scores reach 88.2, indicating strong proficiency in software development workflows.

Reasoning capabilities have also been significantly improved. On the SWE-bench dataset, Mistral Medium 3 completes 65% of tasks successfully, a 12% improvement over Mistral Medium 2. These metrics suggest that the model is not only faster but also more reliable for complex, multi-step logical operations required in enterprise applications.

  • MMLU: 85.4
  • HumanEval: 88.2
  • SWE-bench: 65% Success Rate
  • Latency: <50ms on A100

API Pricing

Mistral AI has structured its API pricing to remain competitive with other top-tier providers. Developers can expect input costs of $0.20 per million tokens and output costs of $0.60 per million tokens. This pricing model is significantly lower than many proprietary equivalents, making it viable for high-volume applications.

A generous free tier is also available for new users, allowing for 100,000 tokens per month at no cost. This tier is sufficient for prototyping and testing, ensuring that developers can evaluate the model's capabilities before committing to a paid plan. The pricing structure is designed to scale linearly with usage, providing predictable costs for production workloads.

  • Input Cost: $0.20 / 1M tokens
  • Output Cost: $0.60 / 1M tokens
  • Free Tier: 100K tokens/month
  • No hidden fees

Comparison Table

When compared to direct competitors, Mistral Medium 3 offers a unique balance of cost and capability. While GPT-4o offers superior multimodal integration, Mistral Medium 3 provides a more open and cost-effective alternative for text-heavy and code-centric tasks. The table below summarizes the key differentiators across the market.

  • Competitive pricing
  • Open weights
  • Edge deployment ready

Use Cases

Mistral Medium 3 is best suited for applications requiring high reasoning capabilities without the overhead of massive parameter counts. Software development teams can utilize the model for code generation and debugging, leveraging its strong HumanEval scores. Additionally, its multilingual support makes it ideal for customer support bots and global content generation platforms.

For researchers and data scientists, the Apache 2.0 license allows for fine-tuning on proprietary datasets without legal concerns. The model's ability to run on single GPUs via the Ministral variants enables distributed intelligence setups, where multiple edge devices collaborate to solve complex problems.

  • Software Development & Coding
  • Multilingual Customer Support
  • Edge AI Deployment
  • RAG Systems & Fine-Tuning

Getting Started

Accessing Mistral Medium 3 is straightforward for developers familiar with standard API integrations. The model is available via the Mistral AI API endpoint and can be accessed using their official Python SDK. For those preferring local deployment, weights are hosted on Hugging Face under the Apache 2.0 license.

To begin, register for an API key on the Mistral platform. The SDK provides examples for chat completion and embedding generation. Documentation includes comprehensive guides on optimizing inference for edge devices, ensuring that teams can deploy the model efficiently across their infrastructure.

  • API Endpoint: api.mistral.ai
  • SDK: Python, Node.js
  • Weights: Hugging Face
  • Docs: mistral.ai/docs

Comparison

API Pricing β€” Input: $0.20 / Output: $0.60 / Context: 128K


Sources

Mistral Closes In on Big AI Rivals

Inside Mistral 3's Big Return

Mistral AI Partners with Nvidia