Skip to content
Back to Blog
Model Releases

StableLM 2: Stability AI's New Open-Source LLMs Challenge Industry Giants

Stability AI releases StableLM 2 with 1.6B and 12B parameters, trained on 2 trillion tokens, delivering competitive performance against larger models.

February 6, 2024
Model ReleaseStableLM 2
StableLM 2 - official image

Introduction

Stability AI has made waves in the open-source community with the February 6, 2024 release of StableLM 2, a new family of open language models that promises to deliver exceptional performance while maintaining accessibility through open-source licensing. This release represents a significant milestone in making powerful AI models available to developers without the constraints of proprietary systems.

What sets StableLM 2 apart is its ability to compete with much larger models despite having a significantly smaller parameter count. The 12B variant achieves performance metrics that rival models like Mistral-7B, while the lightweight 1.6B version offers an attractive option for resource-constrained environments.

Built on Stability AI's commitment to democratizing AI technology, these models represent the next evolution in their open-source language modeling efforts. The release comes with comprehensive training on diverse datasets totaling 2 trillion tokens, ensuring robust performance across multiple domains.

For developers and researchers seeking alternatives to closed-source options, StableLM 2 provides both technical excellence and legal clarity through the Stability AI Community License, making it suitable for commercial deployment scenarios.

Key Features & Architecture

StableLM 2 comes in two distinct configurations: a compact 1.6 billion parameter model and a more capable 12 billion parameter variant. Both models utilize transformer architecture optimized for efficient inference while maintaining high-quality outputs across various tasks.

The training process involved processing approximately 2 trillion tokens from diverse sources including Falcon RefinedWeb, RedPajama, The Pile, and CulturaX datasets. This extensive training corpus ensures the models demonstrate strong performance across multiple languages and domains.

While specific architectural innovations like Mixture of Experts (MoE) implementations haven't been explicitly detailed, the models show impressive efficiency metrics that suggest thoughtful architectural choices. The context window specifications remain competitive with industry standards.

Both variants support standard transformer operations and can be deployed across various hardware configurations, from consumer GPUs to enterprise infrastructure. The models are designed for flexibility in deployment scenarios ranging from edge devices to cloud-based services.

  • Available in 1.6B and 12B parameter configurations
  • Trained on 2 trillion tokens from diverse datasets
  • Transformer architecture with optimization for efficiency
  • Standard context windows suitable for most applications

Performance & Benchmarks

StableLM 2 demonstrates remarkable performance considering its relatively modest parameter count. The 12B variant achieves MMLU scores competitive with models like Mistral-7B, often exceeding expectations for its size category. On coding benchmarks, the model shows strong performance on HumanEval and related assessments.

In reasoning tasks, StableLM 2 maintains consistency across various evaluation frameworks, showing particular strength in mathematical and logical problem-solving scenarios. The SWE-bench evaluations indicate solid capabilities in software engineering tasks.

Compared to previous StableLM iterations, version 2 shows substantial improvements in coherence, factual accuracy, and task completion rates. The training methodology refinements contribute to better instruction following and reduced hallucination rates.

Despite being significantly smaller than many competing models, StableLM 2's 12B variant delivers performance that justifies its adoption in production environments where larger models might be overkill or cost-prohibitive.

  • Competitive MMLU scores vs. larger models like Mistral-7B
  • Strong performance on HumanEval and coding benchmarks
  • Improved coherence and reduced hallucination vs. predecessors
  • Solid SWE-bench results for software engineering tasks

API Pricing

StableLM 2 follows Stability AI's approach to accessible pricing, though specific commercial API costs have not been standardized yet since these are primarily open-source models meant for self-hosting. The open-source nature means no per-token fees when running locally.

For cloud-hosted implementations, pricing varies by provider, but typically ranges around $0.10-$0.20 per million input tokens and $0.20-$0.40 per million output tokens for managed services hosting StableLM 2 models.

Self-hosting eliminates recurring token-based costs entirely, making it particularly attractive for high-volume use cases. The 1.6B model enables cost-effective deployment on consumer hardware, while the 12B variant requires more substantial resources.

The Stability AI Community License allows for commercial use with appropriate attribution, providing clear pathways for business integration without complex licensing negotiations.

  • Primarily open-source for self-hosting (no token fees)
  • Managed service estimates: $0.10-0.20M input, $0.20-0.40M output
  • Self-hosting eliminates per-token costs completely
  • Commercial use permitted under Stability AI Community License

Comparison Table

When comparing StableLM 2 against similar models, several factors become apparent regarding performance, cost, and suitability for different applications.

The table below illustrates how StableLM 2 positions itself relative to other prominent open-source models in terms of key specifications and capabilities.

Cost considerations vary significantly depending on deployment method, with self-hosting offering the best economics for high-volume usage.

Performance-to-size ratios favor StableLM 2's 12B model compared to competitors of similar or larger sizes.

Use Cases

StableLM 2 excels in coding assistance applications, with the 12B variant showing particular strength in code generation, debugging, and documentation tasks. The model's training on diverse programming languages makes it suitable for multi-language development environments.

For conversational AI and chatbot applications, both models provide solid foundation capabilities, with the 12B version offering more nuanced responses and better contextual understanding. The 1.6B model serves well for simpler query-response scenarios.

Reasoning and analytical tasks benefit from StableLM 2's improved training methodology, making it suitable for educational tools, research assistance, and decision support systems.

RAG (Retrieval-Augmented Generation) implementations work particularly well with these models due to their strong comprehension abilities and consistent output quality across extended interactions.

  • Code generation and debugging assistance
  • Conversational AI and chatbot foundations
  • Educational tools and research assistance
  • RAG implementations and document processing

Getting Started

Accessing StableLM 2 begins with visiting Hugging Face or Stability AI's official repositories, where both 1.6B and 12B variants are available for download. The models come with comprehensive documentation covering installation and basic usage patterns.

For local deployment, ensure your system meets the memory requirements: approximately 4GB for the 1.6B model and 24GB for the 12B variant during inference. Various quantization options help reduce memory footprint for resource-constrained environments.

Integration with existing frameworks like Transformers, vLLM, and Text Generation WebUI provides familiar interfaces for developers already working with open-source models.

Community support through Stability AI forums and GitHub discussions helps troubleshoot common implementation challenges and share best practices for optimal performance.

  • Download from Hugging Face or Stability AI repositories
  • 1.6B requires ~4GB, 12B requires ~24GB RAM for inference
  • Compatible with Transformers, vLLM, and Text Generation WebUI
  • Active community support and documentation available

Comparison

API Pricing β€” Input: $0.00 (self-hosted), $0.10-0.20M (managed) / Output: $0.00 (self-hosted), $0.20-0.40M (managed) / Context: Open-source model with commercial use license


Sources

StableLM 2 on Hugging Face

Stability AI Community License