Ministral 3 3B: The Edge AI Revolution Arrives
Mistral AI's newest 3B parameter model brings vision capabilities to local edge devices under Apache 2.0. Perfect for phones and drones.

Introduction
The landscape of artificial intelligence has shifted dramatically in 2025, moving away from massive cloud-dependent models toward efficient, local inference. Mistral AI has capitalized on this trend with the release of Ministral 3 3B on December 2nd, 2025. This new language model is designed specifically for edge computing, enabling developers to deploy sophisticated AI directly onto consumer hardware without relying on expensive cloud APIs. By combining a compact 3B parameter count with advanced vision capabilities, Mistral is democratizing AI for mobile and embedded systems.
This release marks a significant step forward in the industry's push toward artificial general intelligence, as noted by industry analysts tracking the most innovative companies in AI for 2025. Unlike previous iterations that required high-end GPUs, Ministral 3 3B is optimized to run smoothly on standard laptops, smartphones, and even drones. This shift allows for real-time processing of visual data and code generation directly on the device, reducing latency and enhancing user privacy.
- Released: December 2, 2025
- Parameters: 3 Billion
- License: Apache 2.0
- Capabilities: Text + Vision
- Target: Edge Devices
Key Features & Architecture
Ministral 3 3B is engineered with efficiency at its core. The architecture utilizes a dense transformer structure optimized for low-latency inference, ensuring that complex reasoning tasks do not bog down mobile processors. It features a native multimodal input layer that allows the model to process images alongside text prompts seamlessly. This vision capability is crucial for applications like drone navigation or mobile document analysis, where the model must understand visual context instantly.
The model supports a context window of 128k tokens, allowing it to handle extensive documents or video transcripts without losing coherence. Furthermore, the Apache 2.0 license ensures that developers can modify, distribute, and commercialize the model without restrictive clauses. This open-source approach fosters a vibrant ecosystem where community improvements can be rapidly integrated into the core model weights.
- Context Window: 128k tokens
- Input Types: Text + Vision
- License: Apache 2.0
- Quantization Support: INT4, FP4
- Inference Speed: 40 tokens/sec on CPU
Performance & Benchmarks
Despite its small footprint, Ministral 3 3B delivers competitive performance metrics compared to larger models. In the MMLU benchmark, it scores 68.5%, which is a 5% improvement over the previous 3B parameter baseline. On HumanEval, the coding benchmark, it achieves 62% accuracy, demonstrating strong capabilities for local development assistance. These scores are particularly impressive given the model's ability to run entirely on local hardware without external assistance.
In SWE-bench, a complex software engineering benchmark, the model scores 45%, indicating its ability to understand and modify codebases. Compared to competitors like Gemma 2 2B, Ministral 3 3B shows superior reasoning on logic puzzles and mathematical tasks. The vision encoder contributes significantly to these scores, allowing the model to solve problems by analyzing diagrams or code screenshots directly.
- MMLU Score: 68.5%
- HumanEval: 62%
- SWE-bench: 45%
- Vision Accuracy: 85%
- Latency: <100ms on Mobile NPU
API Pricing
While the model weights are free to download under the Apache 2.0 license, Mistral AI also offers a hosted API for developers who prefer managed services. The API pricing is structured to be competitive for high-volume applications. Developers can access the Ministral 3 3B endpoint through the standard Mistral dashboard, with clear pricing tiers for both input and output tokens. This flexibility allows teams to choose between local deployment for cost savings or cloud inference for ease of management.
For the hosted API, the input price is set at $0.20 per million tokens, while the output price is $0.60 per million tokens. Mistral also provides a generous free tier for developers, allowing 100,000 input tokens per month at no cost. This value comparison makes Ministral 3 3B an attractive option for startups and hobbyists looking to build AI agents without incurring significant infrastructure costs.
- Input Price: $0.20 / M tokens
- Output Price: $0.60 / M tokens
- Free Tier: 100k tokens/month
- Payment: Credit Card or Stripe
- Latency SLA: 99.9%
Comparison Table
When evaluating Ministral 3 3B against other leading open-weight models, the trade-offs become clear. While larger models offer higher raw intelligence, Ministral 3 3B wins on efficiency and cost. The comparison below highlights how it stacks up against popular alternatives in terms of context window, output limits, and pricing structure. This data is crucial for architects deciding between cloud-heavy solutions and local edge deployment.
- Efficiency: High
- Cost: Low
- Vision: Native
- License: Apache 2.0
- Deployment: Edge Ready
Use Cases
The versatility of Ministral 3 3B opens up numerous practical applications for developers. In the coding domain, it serves as an excellent local assistant for debugging and refactoring code on personal machines without sending proprietary code to the cloud. For autonomous systems, the vision capabilities make it ideal for drone navigation, where real-time visual processing is critical for safety and efficiency. Additionally, it can be used in customer support agents that need to read and interpret visual data from user inputs.
RAG (Retrieval-Augmented Generation) systems benefit significantly from this model's 128k context window, allowing for long document summarization on local servers. Mobile applications can integrate the model to provide personalized chat experiences that respect user privacy by keeping data on the device. These use cases demonstrate the model's potential to bridge the gap between powerful AI and accessible hardware.
- Local Coding Assistant
- Drone Navigation
- Privacy-Focused Chat
- Mobile RAG Systems
- Edge Vision Analysis
Getting Started
Accessing Ministral 3 3B is straightforward for developers. You can download the weights directly from the Hugging Face repository or use the Mistral AI SDK for Python. For local deployment, ensure your hardware meets the minimum requirements, which include 8GB of RAM for INT4 quantization. The Mistral API documentation provides detailed guides on setting up endpoints, managing keys, and integrating the model into your existing applications.
To start using the API, sign up for a Mistral account and generate an API key. You can then send POST requests to the model endpoint with your text or image payloads. The SDK handles tokenization and response parsing automatically, simplifying the integration process. For local users, the provided Docker containers allow for one-click deployment, making it easy to test the model's performance on your specific hardware configuration.
- Platform: Hugging Face
- SDK: Python, Node.js
- Container: Docker
- Docs: mistral.ai/docs
- Community: Discord
Comparison
API Pricing β Input: 0.20 / Output: 0.60 / Context: 128k