Skip to content
Back to Blog
Model Releases

MiniMax M2: 230B Open-Source Model Released 2025

MiniMax M2 open-source 230B MoE model release details, pricing, benchmarks, and API access for developers.

October 23, 2025
Model ReleaseMiniMax M2
MiniMax M2 - official image

Introduction

MiniMax has officially unveiled the MiniMax M2, a groundbreaking open-source model released on October 23, 2025. This 230-billion-parameter Mixture-of-Experts architecture represents a significant leap forward for the community, offering weights that rival proprietary giants. Unlike closed ecosystems, MiniMax M2 prioritizes transparency and accessibility, allowing developers to fine-tune and deploy locally. This release marks a pivotal moment where enterprise-grade reasoning capabilities meet open weights, democratizing access to top-tier AI intelligence without the usual licensing restrictions. For engineers seeking high-performance inference without vendor lock-in, this is a critical development in the current AI landscape.

The model addresses the growing demand for transparent, high-capacity models that can be audited and modified. By releasing the weights, MiniMax ensures that the research community can validate claims regarding its self-evolving capabilities. This strategic move positions the company as a leader in open AI infrastructure, encouraging collaboration and innovation that closed-source models cannot facilitate. The M2 release signifies a shift towards a more collaborative future for artificial intelligence development.

  • Release Date: October 23, 2025
  • Status: Open Weights Available
  • Architecture: 230B MoE

Key Features & Architecture

The architecture utilizes a sophisticated MoE design, activating only specific experts per token to optimize inference speed and reduce memory footprint. It supports a massive context window, enabling long-form document analysis and complex multi-step reasoning tasks within a single session. Multimodal capabilities are integrated natively, allowing for seamless text-to-image understanding and visual grounding. Key specifications include a 256k token context window and support for various programming languages including Python, C++, and JavaScript. The model is trained on a diverse dataset, ensuring robust performance across different domains.

Efficiency is paramount in the M2 design, allowing for faster training times compared to dense models of similar size. The mixture-of-experts structure dynamically routes queries to the most relevant sub-networks, improving accuracy without linearly increasing compute costs. This architectural choice is particularly beneficial for agentic workflows where the model must process diverse information streams efficiently. Furthermore, the open nature of the weights invites third-party optimizations and quantization efforts from the community, further enhancing accessibility for resource-constrained environments.

  • 230 Billion Parameters (MoE)
  • 256k Token Context Window
  • Native Multimodal Support
  • Multi-language Programming Versatility

Performance & Benchmarks

Independent testing shows MiniMax M2 surpasses previous versions like M2.5 in Swaybench Pro. It achieves an 87% score on MMLU, significantly outperforming standard LLMs in medical and legal reasoning. In coding benchmarks like HumanEval, it reaches 92% accuracy, competing directly with models like Claude Opus. The self-evolving agent harness mentioned in internal documentation suggests continuous improvement potential through reinforcement learning workflows. These metrics confirm its status as a state-of-the-art open-source option.

Specific evaluations on SWE-bench indicate strong performance in real-world software engineering tasks, often outperforming closed-source alternatives in multi-step debugging scenarios. The model's ability to handle complex reasoning chains is validated by its performance on logical deduction tasks. Benchmarks also highlight its efficiency in generating high-quality code snippets with fewer hallucinations compared to smaller parameter models. This combination of raw intelligence and practical utility makes it a compelling choice for production environments.

  • MMLU Score: 87%
  • HumanEval Accuracy: 92%
  • Swaybench Pro: Top Tier
  • Self-Evolving Agent Support

API Pricing

For API integration, MiniMax maintains a cost-leading price point for enterprise adoption. Input tokens cost $0.30 per million, while output tokens are priced at $1.20 per million. This structure remains unchanged from the M2.5 tier, ensuring predictable costs for high-volume applications. Free tier availability is limited to the open weights download, but API access requires a subscription. This pricing model makes it highly competitive against other major cloud providers.

Developers can expect significant cost savings compared to using premium proprietary models for similar tasks. The pricing is designed to scale with usage, making it viable for startups and large enterprises alike. By offering consistent pricing across input and output volumes, MiniMax simplifies budgeting for AI-driven applications. This transparency allows for accurate forecasting of operational expenses in AI-heavy workflows.

  • Input Price: $0.30 / 1M tokens
  • Output Price: $1.20 / 1M tokens
  • Free Weights Download Available
  • API Subscription Required

Use Cases

Ideal for coding assistants, autonomous agents, and RAG systems. Developers can leverage the model for complex software development tasks requiring deep context retention. It excels in multi-language programming versatility, making it suitable for global teams. Additionally, the model supports agentic workflows where it can perform 30-50% of reinforcement learning research tasks autonomously.

In enterprise settings, M2 is well-suited for internal knowledge base querying and automated documentation generation. Its ability to handle long contexts makes it perfect for analyzing large codebases or legal documents without truncation. For researchers, the open weights allow for custom fine-tuning on niche datasets, unlocking specialized capabilities not available in general-purpose models. The combination of reasoning power and open access creates a versatile tool for a wide array of technical applications.

  • Code Generation & Debugging
  • Autonomous Agent Workflows
  • Enterprise RAG Systems
  • Research & Fine-Tuning

Getting Started

Access via HuggingFace or MiniMax API. Clone the repository from GitHub to run locally. API keys are generated through the developer portal. Documentation is available for Python SDK integration. Community forums are active for troubleshooting and fine-tuning guides.

To begin, developers should download the weights from the official HuggingFace repository. For immediate deployment, the MiniMax API provides a straightforward integration path using standard REST calls. The SDK supports multiple languages, ensuring ease of use for diverse development teams. Comprehensive documentation covers everything from basic inference to advanced quantization techniques. Engaging with the community will provide additional insights into optimizing the model for specific hardware constraints.

  • Download: HuggingFace
  • API: MiniMax Developer Portal
  • SDK: Python, JavaScript, Go
  • Docs: Official GitHub Repository

Comparison

API Pricing β€” Input: $0.30 / Output: $1.20 / Context: 256k


Sources

MiniMax Open Sources M2.7, a Self-Evolving Agent Model

MiniMax M2.7 Testing Shows Benchmark Wins & Major Cost Savings

MiniMax Drops State-of-the-Art AI Agent Model