BLOOMZ is the instruction-tuned variant of BLOOM, fine-tuned on multilingual tasks to follow user instructions across 46 languages. It represents a massive multilingual instruction-following model from the BigScience collaboration.
- Parameters: 176 billion (same as BLOOM)
- Languages: 46 natural languages
- Programming Languages: 13
- Base: BLOOM 176B
- Training: Fine-tuned on xP3 dataset
Multilingual Instruction-Following:
- Instruction-tuning across 46 languages
- Cross-lingual task generalization
- Multilingual zero-shot capabilities
- Global accessibility
- Full 176B parameter model
- Complete instruction-tuning
- All 46 languages
- Smaller 7.1B variant
- More efficient deployment
- Maintained multilingual capabilities
- Compact 3B variant
- Consumer hardware friendly
- Core language support
- Smallest 1.7B variant
- Edge deployment capable
- Key languages maintained
Cross-lingual Public Pool of Prompts (xP3):
- Multilingual instruction dataset
- Diverse task coverage
- 46 language support
- Cross-lingual examples
- Task instructions and demonstrations
- Massive Multilingual: 46 natural languages
- Instruction-Following: User command execution
- Cross-Lingual: Transfer across languages
- Open Source: Freely available
- Multiple Sizes: 1.7B to 176B parameters
- Code Support: 13 programming languages
Includes major languages from:
- Europe: English, French, Spanish, German, Italian, etc.
- Asia: Chinese, Japanese, Korean, Hindi, Arabic, etc.
- Africa: Multiple African languages
- Americas: Portuguese, indigenous languages
- And many more: 46 total natural languages
Instruction-Following:
- Strong across all 46 languages
- Good cross-lingual transfer
- Competitive multilingual performance
- Maintained BLOOM capabilities
Zero-Shot Tasks:
- Generalization to new instructions
- Cross-lingual task transfer
- Multi-task capabilities
- Worldwide customer support
- Multilingual content generation
- International documentation
- Cross-border communication
- Low-resource language support
- Local language applications
- Regional content creation
- Cultural preservation
- Translation services
- Cross-lingual information retrieval
- Multilingual question answering
- Global knowledge management
- Task execution in any supported language
- Natural language interfaces
- Automated workflows
- Interactive systems
BLOOMZ vs BLOOM:
- BLOOMZ: Instruction-tuned, follows commands
- BLOOM: Base model, text generation
- BLOOMZ: Better task performance
- BLOOM: Raw generation capabilities
- Base: Start with BLOOM 176B
- Dataset: Use xP3 multilingual instructions
- Fine-Tuning: Instruction-following training
- Languages: Maintain all 46 languages
- Evaluation: Cross-lingual testing
- 176B: Maximum capability, requires large infrastructure
- 7B1: Balanced performance and efficiency
- 3B: Consumer hardware deployment
- 1B7: Edge and mobile potential
- Cloud platforms
- On-premises clusters (176B)
- Single GPU (smaller variants)
- Quantization support
Community Achievement:
- Global researcher collaboration
- Multilingual AI advancement
- Open science principles
- Democratized access
- Reproducible research
Architecture: Same as BLOOM (decoder-only transformer) Context Length: Standard for BLOOM Fine-Tuning: xP3 dataset instruction-tuning Inference: Standard transformer inference
- Multilingual: 46 language support
- Instruction-Following: Task execution
- Open Source: Free access
- Multiple Sizes: Flexibility
- Community: Global collaboration
Acknowledged:
- 176B variant requires significant resources
- Language quality varies by corpus size
- Biases from training data
- Not all languages equally represented
BLOOMZ contributed:
- Multilingual instruction-following
- Low-resource language support
- Open-source multilingual AI
- Research democratization
- Global accessibility
- Multilingual NLP research
- Cross-lingual transfer studies
- Instruction-following research
- Low-resource language development
- Bias and fairness studies
Compatible with:
- Hugging Face Transformers
- Standard inference frameworks
- API services
- Custom applications
xP3 Enabled:
- Multilingual instruction-tuning
- Cross-lingual task transfer
- Diverse task coverage
- Open research data
- Continued improvements
- Enhanced language support
- Better cross-lingual transfer
- Community contributions
- Research advancements
BigScience RAIL License - permissive for research and commercial use with responsible AI provisions.
Free and open-source.