Skip to content
Back to Blog
Model Releases

Jamba 1.5: The New Hybrid MoE Standard for Long Context

AI21 Labs launches Jamba 1.5, combining Mamba and Transformer architectures for unprecedented speed and 256K context support in an open-weight release.

August 22, 2024
Model ReleaseJamba 1.5
Jamba 1.5 - official image

Introduction

In the rapidly evolving landscape of large language models, AI21 Labs has officially released Jamba 1.5 on August 22, 2024, marking a significant milestone in hybrid architecture development. This release is not merely an incremental update but a fundamental shift towards efficiency, addressing the critical bottleneck of context window management in enterprise applications. For developers and AI engineers, the availability of open weights alongside a high-performance API means you can now deploy this model without the usual licensing constraints, fostering greater innovation in long-context workflows.

The industry has long struggled with the trade-off between model capacity and inference speed. Jamba 1.5 aims to solve this by integrating the state-space model (SSM) capabilities of Mamba with the attention mechanisms of Transformers. This hybrid approach allows the model to process vast amounts of data without the quadratic complexity associated with standard attention layers, making it the fastest long-context model available at the time of its release.

  • Released Date: August 22, 2024
  • Provider: AI21 Labs
  • License: Open Weights Available
  • Primary Focus: Long-context efficiency and speed

Key Features & Architecture

The core innovation of Jamba 1.5 lies in its unique Mamba-Transformer hybrid MoE (Mixture of Experts) architecture. Unlike pure Transformer models that scale quadratically with context length, Jamba 1.5 utilizes a sparse MoE structure where only a subset of parameters is active during inference. This design choice drastically reduces computational load while maintaining high-quality output generation.

The model boasts a massive total parameter count of 398 billion, with a highly optimized 94 billion active parameters per token. This sparse activation ensures that the model remains lightweight enough for deployment on edge devices or smaller clusters while retaining the intelligence of a much larger model. Additionally, the architecture supports a context window of 256K tokens, enabling the processing of entire books, hours of video transcripts, or massive codebases in a single pass.

  • Total Parameters: 398B
  • Active Parameters: 94B (per token)
  • Context Window: 256K tokens
  • Architecture: Mamba-Transformer Hybrid MoE

Performance & Benchmarks

Performance evaluations show that Jamba 1.5 outperforms its predecessor and direct competitors across standard reasoning and coding benchmarks. The integration of the Mamba state-space model contributes to faster inference speeds, particularly noticeable in long-context scenarios where traditional models often degrade in quality or latency. While specific latency numbers depend on hardware, the throughput improvements are significant enough to enable real-time agent workflows.

In terms of capability, Jamba 1.5 demonstrates superior performance on the MMLU (Massive Multitask Language Understanding) benchmark, scoring higher than standard 405B parameter models. On the HumanEval coding benchmark, it achieves state-of-the-art results for open-weight models, proving its utility for software engineering tasks. Furthermore, on the SWE-bench (Software Engineering benchmark), it shows improved accuracy in solving complex GitHub issues compared to previous iterations.

  • MMLU Score: Improved over Jamba 1.0
  • HumanEval: State-of-the-art for open weights
  • SWE-bench: High accuracy on complex issues
  • Inference Speed: Fastest in class for 256K context

API Pricing

While the model weights are open source, AI21 Labs provides a hosted API for seamless integration into production pipelines. The pricing structure is designed to be competitive for high-volume users while remaining accessible for startups. Developers can access the model via standard REST endpoints or through the AI21 Labs SDK, ensuring compatibility with existing infrastructure.

The API pricing is tiered based on usage volume. For the standard tier, input tokens are priced at $3.50 per million tokens, while output tokens cost $17.50 per million tokens. This pricing reflects the high computational cost of the MoE architecture. However, because the model is open source, users can self-host the weights to avoid these costs entirely, provided they have the necessary GPU resources. A free tier is available for testing purposes, allowing developers to validate performance before committing to paid volumes.

  • Input Cost: $3.50 per million tokens
  • Output Cost: $17.50 per million tokens
  • Free Tier: Available for testing
  • Self-hosting: Supported via open weights

Comparison Table

To contextualize Jamba 1.5's capabilities, we compare it against other leading models in the current market. The comparison highlights its superior context window and competitive pricing structure relative to its performance metrics. This table serves as a quick reference for engineers selecting a model for their specific workload requirements.

  • Comparison includes Jamba 1.5, Llama 3.1 405B, and Mistral Large 2
  • Focus on context window, cost, and key strengths

Use Cases

Jamba 1.5 is uniquely positioned for applications that demand both high reasoning capabilities and extensive context processing. In the realm of coding, it excels at refactoring large codebases and generating documentation from extensive repositories. Its 256K context window makes it ideal for RAG (Retrieval-Augmented Generation) systems that need to ingest entire knowledge bases without chunking data into smaller, potentially context-losing fragments.

For enterprise agents, the speed and efficiency of the Mamba architecture allow for multi-turn conversations without the latency usually associated with long-context models. It is also suitable for legal and financial analysis, where accuracy over long documents is critical. The open-source nature further encourages research into specialized domains, allowing teams to fine-tune the model for niche tasks like medical record analysis or proprietary data processing.

  • Coding & Software Engineering
  • Long-Document Analysis (Legal/Financial)
  • RAG Systems with Large Knowledge Bases
  • Real-time Agent Workflows

Getting Started

Accessing Jamba 1.5 is straightforward for developers familiar with standard LLM integration patterns. You can access the model via the AI21 Labs API endpoint, which supports standard JSON payloads. Alternatively, for local deployment, the weights are available on GitHub under the AI21 Labs organization, allowing you to run the model on your own infrastructure using vLLM or similar inference engines.

To begin, register for an API key on the AI21 Labs dashboard. Then, integrate the SDK into your Python environment to send requests. For local development, clone the repository and follow the provided Docker instructions to spin up an inference container. The documentation includes detailed examples for both chat completion and embedding tasks, ensuring a smooth onboarding process for new users.

  • API Endpoint: api.ai21labs.com
  • SDK: Python, Node.js, Go
  • Weights: GitHub Repository
  • Docs: Official AI21 Labs Documentation

Comparison

API Pricing β€” Input: $3.50 / Output: $17.50 / Context: 256K