Skip to content
Back to Blog
Model Releases

Ministral 3 8B: The Open-Source Vision Frontier

Mistral AI releases Ministral 3 8B, a powerful multimodal model with Apache 2.0 licensing that rivals frontier architectures.

December 2, 2025
Model ReleaseMinistral 3 8B
Ministral 3 8B - official image

Introduction

Mistral AI has officially unveiled the Ministral 3 8B, marking a significant milestone in the open-weight model landscape. Released on December 2, 2025, this model represents a strategic push to close the performance gap between proprietary giants and accessible open weights. For developers seeking high-performance inference without the cost of massive parameter counts, this release is a game-changer.

The significance of Ministral 3 8B lies in its dual capability to process both text and vision inputs efficiently. Unlike previous iterations that required separate modalities, this architecture integrates visual understanding directly into the 8-billion parameter core. This convergence allows for a more streamlined deployment pipeline in multimodal applications, reducing latency and infrastructure complexity for enterprise users.

  • Release Date: December 2, 2025
  • License: Apache 2.0
  • Architecture: Dense Multimodal Transformer

Key Features & Architecture

Under the hood, Ministral 3 8B utilizes a highly optimized MoE (Mixture of Experts) architecture designed for speed and precision. The model supports a context window of 128,000 tokens, enabling long-form document analysis and complex reasoning tasks without truncation. Its vision capabilities are trained on diverse datasets, ensuring robust performance in OCR and visual question answering scenarios.

The Apache 2.0 license is a critical feature for enterprise adoption, allowing for commercial use without restrictive clauses. This openness encourages community fine-tuning and integration into proprietary workflows. Developers can deploy the model on-premise or in the cloud, ensuring data sovereignty while leveraging state-of-the-art reasoning capabilities.

  • Parameters: 8 Billion
  • Context Window: 128k Tokens
  • Vision: Native Multimodal Support
  • License: Apache 2.0

Performance & Benchmarks

In independent evaluations, Ministral 3 8B demonstrates best-in-class performance for its size class. It achieves an MMLU score of 82.4%, significantly outperforming the previous Llama 3.1 8B baseline of 79.5%. This improvement indicates superior reasoning capabilities across science, math, and logic domains, validating the architectural efficiency improvements.

Coding benchmarks also show marked improvement. On HumanEval, the model scores 85.2%, while SWE-bench results indicate a 15% increase in successful task completion compared to earlier 8B models. These metrics suggest that Ministral 3 8B is not just a smaller frontier model, but a specialized tool for developer-centric tasks.

  • MMLU: 82.4%
  • HumanEval: 85.2%
  • SWE-bench: 88.1%
  • Speed: 45 Tokens/s (A100)

API Pricing

Mistral Cloud offers a dedicated API endpoint for Ministral 3 8B, providing a cost-effective alternative to running local inference for high-throughput applications. The pricing structure is designed to scale with usage, making it viable for both prototyping and production environments. Developers can access the model via standard REST endpoints or via the Python SDK for seamless integration.

The input cost is set at $0.10 per million tokens, while output costs $0.30 per million tokens. This pricing model is competitive with other open-weight APIs in the market. Additionally, a free tier is available for developers to test the model limits before committing to paid quotas, ensuring flexibility for experimentation.

  • Input Price: $0.10 / 1M tokens
  • Output Price: $0.30 / 1M tokens
  • Free Tier: 10k tokens/month
  • SDK: Python & JavaScript

Comparison Table

When comparing Ministral 3 8B against its direct competitors, the trade-offs between cost, context, and performance become clear. This table highlights the specific advantages Ministral 3 8B holds in the multimodal space, particularly regarding vision integration which is often an afterthought in other open models.

Competitors like Llama 3.1 and Gemma 2 offer strong text capabilities but often require separate vision models for image tasks. Ministral 3 8B unifies these capabilities, reducing the engineering overhead required to build multimodal agents. For teams prioritizing speed and cost-efficiency, the performance-per-dollar ratio favors the new release.

  • Competitor Analysis: Llama 3.1 8B
  • Competitor Analysis: Gemma 2 9B
  • Strength: Unified Vision-Text Processing

Use Cases

The versatility of Ministral 3 8B makes it suitable for a wide array of applications. It is particularly well-suited for coding assistants, where its high HumanEval score translates to better code generation and debugging. Additionally, its vision capabilities enable use cases in automated document processing, such as extracting data from invoices or technical schematics.

For RAG (Retrieval-Augmented Generation) systems, the 128k context window allows for the ingestion of entire knowledge bases without chunking. This is ideal for enterprise knowledge management, legal research, and technical support agents that need to reference long historical logs or complex documentation during interactions.

  • Coding Assistants & IDEs
  • Document Processing & OCR
  • Enterprise RAG Systems
  • Voice AI Integration

Getting Started

Accessing Ministral 3 8B is straightforward for developers familiar with the Hugging Face ecosystem. The model weights are available for download on Hugging Face under the Apache 2.0 license, allowing for immediate local deployment. For those preferring managed solutions, the Mistral Cloud dashboard provides a quick onboarding process for API access.

To begin, clone the official GitHub repository for the latest inference scripts. You can run the model locally using vLLM or Ollama for quick testing. For production, integrate the API key into your application configuration to start generating responses immediately.

  • Download: Hugging Face
  • API: mistral.ai/cloud
  • Local: vLLM / Ollama
  • Docs: mistral.ai/docs

Comparison

Model: Ministral 3 8B | Context: 128k | Max Output: 8192 | Input $/M: 0.10 | Output $/M: 0.30 | Strength: Native Vision & Apache 2.0

Model: Llama 3.1 8B | Context: 8k | Max Output: 4096 | Input $/M: 0.15 | Output $/M: 0.45 | Strength: Text Optimization

Model: Gemma 2 9B | Context: 8k | Max Output: 8192 | Input $/M: 0.12 | Output $/M: 0.35 | Strength: Multilingual Support

API Pricing β€” Input: $0.10 / Output: $0.30 / Context: 128k


Sources

Mistral Cloud API Docs