BLOOM
BigScience 176B multilingual causal LM—landmark collaborative open training effort on Jean Zay (weights under BigScience Responsible AI License).
Why it is included
Historical reference for multilingual open LLM governance and collective training.
Best for
Policy papers, multilingual baselines, and archival research.
Strengths
- Multilingual breadth
- Open collaboration precedent
Limitations
- Huge to run; newer models often preferred for efficiency
Good alternatives
OLMo · Llama · Mistral
Related tools
AI & Machine Learning
Hugging Face Transformers
State-of-the-art pretrained models for PyTorch, TensorFlow, and JAX.
AI & Machine Learning
OLMo
Allen AI fully open LLM **pipeline**: weights, training code, data mixes, and evaluation—research transparency flagship.
AI & Machine Learning
Qwen
Alibaba’s Qwen family (dense and MoE) with strong multilingual and coding variants; weights and code on Hugging Face under stated licenses per release.
AI & Machine Learning
Yi
01.AI Yi open-weight bilingual models (EN/ZH focus) with Apache-2.0 or Yi license per checkpoint on Hugging Face.
AI & Machine Learning
Meta Llama (open models)
Meta’s Llama family of open **weights** (subject to Llama license) with reference code, tooling, and downloads via Hugging Face and meta-llama org.
AI & Machine Learning
Mistral AI (open models)
Mistral’s open-weight checkpoints (e.g. 7B era, Mixtral MoE) and Apache-2.0–licensed **code** alongside proprietary flagship lines—verify each checkpoint.
