Skip to content

Latest commit

 

History

History
47 lines (30 loc) · 1.26 KB

File metadata and controls

47 lines (30 loc) · 1.26 KB

Overview

StableLM is a family of efficient language models from Stability AI designed for practical deployment. The models prioritize real-world utility and efficiency over pure parameter count.

Model Variants

  • StableLM 1.6B: Trained on 2 trillion tokens
  • StableLM 3B: Mid-size variant
  • StableLM 7B: Larger variant with enhanced capabilities

Architecture

  • Efficient transformer architecture
  • Optimized for inference speed
  • Designed for deployment on modest hardware

Performance

The 1.6B model beats other sub-2B options despite its compact size, demonstrating efficient training and architecture design.

Key Features

  • Efficiency: Optimized for fast inference
  • Practical Focus: Built for developers who need working code
  • Extensive Training: 2 trillion tokens for the 1.6B model
  • Commercial Use: Available for commercial applications
  • Easy Deployment: Runs on consumer hardware

Use Cases

  • Rapid prototyping
  • Edge deployment
  • Resource-constrained environments
  • Local development
  • Cost-effective production deployment

Training Data

Trained on diverse, high-quality data with emphasis on code and technical content.

Licensing

Available under permissive open-source license.

Pricing

Free and open-source.