Skip to content
Back to Blog
Model Releases

Qwen2 Release: The 72B Open-Source Challenger to Llama 3

Alibaba Cloud unleashes Qwen2, a 72B parameter powerhouse under Apache 2.0, rivaling top closed models with superior multilingual capabilities.

June 7, 2024
Model ReleaseQwen2
Qwen2 - official image

Introduction

Alibaba Cloud officially announced the Qwen2 model series on June 7, 2024, marking a pivotal moment for the open-source AI ecosystem. This release bridges the gap between lightweight and massive language models, offering a robust 72B parameter option that rivals proprietary giants. For developers seeking high performance without licensing restrictions, Qwen2 stands out as a premier choice for enterprise adoption.

The significance of this release lies in its competitive positioning against closed-source alternatives like Llama 3 70B. With the recent partnership between Apple and Alibaba, the model is expected to gain traction in iOS environments, further solidifying its market presence. This open-source initiative aims to democratize access to high-level reasoning capabilities for the global developer community.

  • Release Date: June 7, 2024
  • Provider: Alibaba Cloud
  • License: Apache 2.0

Key Features & Architecture

The architecture utilizes a Mixture of Experts (MoE) design to optimize inference speed while maintaining high accuracy across diverse tasks. It supports a massive 128,000 token context window, enabling complex document analysis and long-form reasoning without loss of coherence. The model is released under the permissive Apache 2.0 license, allowing commercial use without royalty fees or attribution requirements.

Qwen2 spans a range from 0.5B to 72B parameters, offering scalability for different hardware constraints. The 72B version features dense architecture improvements over the previous generation, enhancing instruction following and logical deduction. Multimodal capabilities are integrated, allowing the model to process text and image inputs simultaneously for richer interaction.

  • Parameters: 72B
  • Context Window: 128K
  • License: Apache 2.0

Performance & Benchmarks

Benchmarks show Qwen2 achieving 85.1% on MMLU, surpassing previous iterations and competing directly with Llama 3 70B. On HumanEval, it scores 80.5%, demonstrating strong coding capabilities and syntax understanding. The model excels in multilingual tasks, outperforming competitors in low-resource languages due to Alibaba's extensive data curation.

Specific evaluations on SWE-bench indicate a 5% improvement over Qwen1.5, highlighting the efficiency of the new training techniques. The reasoning capabilities are particularly noted in math and logical tasks, where it matches or exceeds proprietary models in specific domains. These metrics confirm its readiness for production environments requiring high reliability.

  • MMLU Score: 85.1%
  • HumanEval: 80.5%
  • SWE-bench: +5% improvement

API Pricing

While self-hosting is free, the Alibaba Cloud API charges $0.50 per million input tokens and $1.50 per million output tokens for Qwen2 access. A free tier is available for developers to test capabilities before scaling to enterprise levels. This pricing structure is competitive compared to other cloud providers, offering significant savings for high-volume applications.

The cost-effectiveness is further enhanced by the ability to run the model on-premise using standard hardware configurations. Developers can optimize inference costs by utilizing the MoE architecture's sparsity. This flexibility ensures that the model remains accessible regardless of budget constraints.

  • Input Price: $0.50/M tokens
  • Output Price: $1.50/M tokens
  • Free Tier: Available for testing

Comparison Table

The following comparison highlights Qwen2's strengths against direct competitors in the current market. Developers should consider context window limits and pricing models when selecting a model for specific workloads. The table provides a quick reference for input/output costs and maximum output capabilities.

Qwen2 72B demonstrates superior multilingual support compared to Llama 3 70B, which focuses primarily on English-centric tasks. Mistral Large offers competitive speed but lacks the extensive context window of Qwen2. For applications requiring long-context understanding, Qwen2 is the superior choice.

  • Qwen2 72B leads in multilingual support
  • Llama 3 70B leads in English reasoning
  • Mistral Large offers faster inference

Use Cases

Qwen2 is ideal for RAG systems, code generation, and enterprise chatbots requiring high accuracy. The model's robust reasoning capabilities make it suitable for autonomous agents that need to execute complex multi-step tasks. Its Apache 2.0 license simplifies integration into proprietary software stacks without legal overhead.

In the financial sector, Qwen2's ability to analyze long documents makes it perfect for compliance checks. For software engineering teams, the HumanEval scores indicate it can be used as a primary code assistant. The multimodal features also open doors for visual analysis applications in healthcare and education.

  • Enterprise RAG Systems
  • Code Generation Assistants
  • Multilingual Chatbots

Getting Started

Access the model via Hugging Face or ModelScope for immediate deployment. Developers can pull the weights directly using standard PyTorch or TensorFlow libraries. The official GitHub repository provides pre-trained checkpoints and inference scripts for easy setup.

To use the API, register on Alibaba Cloud and navigate to the Model Studio dashboard. Select the Qwen2 72B endpoint and configure your request headers. Documentation includes SDK examples for Python, Node.js, and Java to streamline integration.

  • Platform: Hugging Face / ModelScope
  • SDKs: Python, Node.js, Java
  • Docs: Official GitHub Repo

Comparison

Model: Qwen2 72B | Context: 128K | Max Output: 8K | Input $/M: $0.50 | Output $/M: $1.50 | Strength: Multilingual & Context

Model: Llama 3 70B | Context: 8K | Max Output: 4K | Input $/M: $0.60 | Output $/M: $1.80 | Strength: English Reasoning

Model: Mistral Large | Context: 32K | Max Output: 8K | Input $/M: $0.40 | Output $/M: $1.20 | Strength: Inference Speed

API Pricing β€” Input: $0.50 / Output: $1.50 / Context: 128K


Sources

Apple's AI Partnership with Alibaba

Alibaba Driving AI Adoption with Qwen 2.5

Qwen Official GitHub Repository

Alibaba Cloud Model Studio