RWKV
RNN-meets-transformer linear-attention LM architecture running with O(n) memory—unique open line for long-context and embedded inference.
Why it is included
Non-transformer open architecture with active community ports (CUDA, Metal, Web).
Best for
Experimenters wanting recurrent-style LLMs without full KV cache growth.
Strengths
- Linear time
- Embedded friendly
- Multiple runtimes
Limitations
- Ecosystem smaller than Llama for tooling
Good alternatives
Transformer LMs (Llama-class) · Linear-attention research stacks
Related tools
AI & Machine Learning
llama.cpp
Plain C/C++ inference for LLaMA-class models with broad community backends.
AI & Machine Learning
PyTorch
Deep learning framework with strong research-to-production paths.
AI & Machine Learning
Meta Llama (open models)
Meta’s Llama family of open **weights** (subject to Llama license) with reference code, tooling, and downloads via Hugging Face and meta-llama org.
AI & Machine Learning
Mistral AI (open models)
Mistral’s open-weight checkpoints (e.g. 7B era, Mixtral MoE) and Apache-2.0–licensed **code** alongside proprietary flagship lines—verify each checkpoint.
AI & Machine Learning
Qwen
Alibaba’s Qwen family (dense and MoE) with strong multilingual and coding variants; weights and code on Hugging Face under stated licenses per release.
AI & Machine Learning
DeepSeek
DeepSeek open-weight models (e.g. V3/R1 lineage) with MIT or custom terms per release—high capability coding and reasoning checkpoints.
