Skip to content
Back to Blog
Model Releases

MiniMax M2.5: The Open-Source Frontier Coding Model for 2026

MiniMax releases M2.5, a 230B MoE model achieving 80.2% on SWE-Bench Verified with aggressive cost-efficiency for AI agents.

February 12, 2026
Model ReleaseMiniMax M2.5
MiniMax M2.5 - official image

Introduction

On February 12, 2026, MiniMax officially unveiled the MiniMax M2.5, a groundbreaking frontier model designed specifically for high-performance coding tasks. This release signifies a major shift in the open-source AI landscape, offering developers a powerful alternative to proprietary giants without the prohibitive costs associated with them. The model is engineered to handle complex software engineering workflows, from writing boilerplate code to debugging intricate systems.

What makes the M2.5 particularly notable is its commitment to accessibility. Unlike many closed models that lock away their weights, MiniMax has fully open-sourced the M2.5 architecture. This allows the community to fine-tune, deploy, and integrate the model into custom pipelines. The primary goal of this release is to democratize access to frontier-level intelligence, ensuring that startups and individual developers can leverage state-of-the-art capabilities without burning through their budgets.

  • Released Date: 2026-02-12
  • Provider: MiniMax
  • License: Open Source

Key Features & Architecture

The M2.5 utilizes a sophisticated Mixture-of-Experts (MoE) architecture that balances raw parameter count with computational efficiency. While the model boasts a total of 230 billion parameters, it employs a sparse activation mechanism that only activates 10 billion parameters per token. This design choice significantly reduces the computational load during inference, allowing for faster processing speeds and lower energy consumption compared to dense models of similar size.

This sparse activation strategy is crucial for enabling always-on AI agents that require continuous operation without incurring massive cloud costs. The model supports a context window of 128K tokens, enabling it to process extensive documentation and long-form codebases in a single pass. Furthermore, the M2.5 is optimized for multimodal inputs, allowing it to interpret diagrams and text simultaneously, which is vital for modern software development environments.

  • Total Parameters: 230B
  • Active Parameters: 10B per token
  • Architecture: MoE (Mixture of Experts)

Performance & Benchmarks

In terms of raw capability, the MiniMax M2.5 has set new standards for coding benchmarks. It achieved an impressive score of 80.2% on the SWE-Bench Verified dataset, a metric that measures the ability to solve real-world GitHub issues. This score places it in direct competition with top-tier proprietary models like GPT-4o and Claude 3.5 Sonnet. The model excels not just in passing tests, but in understanding the nuances of legacy code and applying best practices automatically.

Beyond coding, the M2.5 demonstrates strong agentic capabilities, meaning it can plan and execute multi-step tasks autonomously. Benchmarks show that the model can handle reinforcement learning workflows more effectively than previous iterations. This makes it ideal for applications requiring iterative improvement and self-correction. The performance gains are attributed to advanced reinforcement learning techniques used during the training phase, which fine-tuned the model specifically for reasoning under constraints.

  • SWE-Bench Verified: 80.2%
  • HumanEval: Top Tier
  • Agentic Capability: High

API Pricing

One of the most compelling aspects of the MiniMax M2.5 is its pricing structure, which is designed to be accessible for both hobbyists and large-scale enterprises. The input cost is set at $0.30 per million tokens, while the output cost is $1.20 per million tokens. This pricing model is approximately one-tenth the cost of comparable proprietary solutions, making it economically viable for high-volume applications.

MiniMax has also targeted a specific cost-efficiency metric for inference. They claim a processing cost of just $1 per hour for 100 tokens per second throughput. This efficiency is driven by the sparse activation mechanism mentioned earlier. Additionally, a free tier is available for developers to test the API limits before committing to a paid plan, ensuring that the transition to production is smooth and low-risk.

  • Input Price: $0.30 / 1M tokens
  • Output Price: $1.20 / 1M tokens
  • Free Tier: Available for testing

Comparison Table

When comparing the MiniMax M2.5 against industry leaders, the trade-offs become clear. While GPT-4o offers a robust ecosystem, the M2.5 offers superior cost efficiency for open-source deployments. Claude 3.5 Sonnet provides a larger context window, but the M2.5 matches its reasoning capabilities at a fraction of the price. For developers prioritizing budget and open-source flexibility, M2.5 is the clear winner.

  • Cost Efficiency: 10x cheaper than GPT-4o
  • Open Weights: Yes
  • SWE-Bench: 80.2%

Use Cases

The MiniMax M2.5 is best suited for a variety of advanced applications. In the realm of coding, it serves as an autonomous agent capable of refactoring legacy codebases and generating unit tests. For reasoning tasks, it can solve complex mathematical problems and logical puzzles that require multi-step deduction. Additionally, its multimodal capabilities make it excellent for RAG (Retrieval-Augmented Generation) systems that need to query technical documentation alongside code repositories.

Developers can also use the model to build self-evolving software agents. Since the model is open source, researchers can modify the weights to adapt it to niche programming languages or specific industry standards. This flexibility is unmatched by closed models and positions the M2.5 as a foundational tool for the next generation of AI-powered development tools.

  • Autonomous Coding Agents
  • Legacy Code Refactoring
  • RAG Pipelines

Getting Started

Accessing the MiniMax M2.5 is straightforward for developers. The model is available via a public API endpoint that supports standard REST protocols. MiniMax also provides official SDKs for Python and JavaScript, simplifying the integration process for common development frameworks. For those who prefer local deployment, the model weights are hosted on GitHub, allowing for on-premise inference using compatible hardware accelerators.

To get started, developers should register for an API key on the MiniMax developer portal. Documentation is comprehensive, including examples for chat completion and function calling. The community is also encouraged to contribute to the open-source repository to help improve the model's performance on edge cases. With these resources, the barrier to entry for using frontier AI is significantly lowered.

  • API Endpoint: Available via Portal
  • SDKs: Python, JavaScript
  • GitHub: Weights Open Source

Comparison

API Pricing β€” Input: 0.30 / Output: 1.20 / Context: 128K


Sources

MiniMax M2.5 Sparse Activation Analysis

MiniMax M2.5 Benchmarks and Pricing

Chinese Start-ups Release New AI Models